modelId
string | author
string | last_modified
timestamp[us, tz=UTC] | downloads
int64 | likes
int64 | library_name
string | tags
list | pipeline_tag
string | createdAt
timestamp[us, tz=UTC] | card
string |
---|---|---|---|---|---|---|---|---|---|
pearsallyhai/blockassist
|
pearsallyhai
| 2025-09-25T06:02:43Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"solitary untamed butterfly",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:15:52Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- solitary untamed butterfly
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
reinaldoadamo57/blockassist
|
reinaldoadamo57
| 2025-09-25T06:02:02Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"galloping miniature warthog",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:15:22Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- galloping miniature warthog
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
daehan-everai/training_output
|
daehan-everai
| 2025-09-25T06:01:57Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"generated_from_trainer",
"trl",
"sft",
"base_model:openai/gpt-oss-120b",
"base_model:finetune:openai/gpt-oss-120b",
"endpoints_compatible",
"region:us"
] | null | 2025-09-24T19:22:14Z |
---
base_model: openai/gpt-oss-120b
library_name: transformers
model_name: training_output
tags:
- generated_from_trainer
- trl
- sft
licence: license
---
# Model Card for training_output
This model is a fine-tuned version of [openai/gpt-oss-120b](https://huggingface.co/openai/gpt-oss-120b).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="daehan-everai/training_output", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/lucas01/llmv3/runs/9grgr93o)
This model was trained with SFT.
### Framework versions
- TRL: 0.23.0
- Transformers: 4.56.2
- Pytorch: 2.8.0+cu128
- Datasets: 4.1.1
- Tokenizers: 0.22.1
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
MAK007/sdxl_seg_controlnet
|
MAK007
| 2025-09-25T06:01:09Z | 0 | 0 |
diffusers
|
[
"diffusers",
"safetensors",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"controlnet",
"diffusers-training",
"base_model:SG161222/RealVisXL_V4.0",
"base_model:adapter:SG161222/RealVisXL_V4.0",
"license:openrail++",
"region:us"
] |
text-to-image
| 2025-09-22T16:31:43Z |
---
base_model: SG161222/RealVisXL_V4.0
library_name: diffusers
license: openrail++
inference: true
tags:
- stable-diffusion-xl
- stable-diffusion-xl-diffusers
- text-to-image
- diffusers
- controlnet
- diffusers-training
---
<!-- This model card has been generated automatically according to the information the training script had access to. You
should probably proofread and complete it, then remove this comment. -->
# controlnet-MAK007/sdxl_seg_controlnet
These are controlnet weights trained on SG161222/RealVisXL_V4.0 with new type of conditioning.
You can find some example images below.
prompt: The image showcases a spacious and well-lit living room with a large window that allows natural light to flood the space. The room features a comfortable couch situated in the center, surrounded by several chairs placed around the room. A dining table is also present, accompanied by a few chairs.In addition to the seating arrangements, the living room is adorned with various decorative elements. There are multiple potted plants placed throughout the room, adding a touch of greenery and life to the space. A vase can be seen on a surface, and a bowl is placed nearby. A clock is mounted on the wall, and a book is resting on a surface, indicating that the room is not only for relaxation but also for reading and other leisure activities.

prompt: The image features a large bathroom with a double sink vanity. The vanity is made of wood and has a marble countertop. The sinks are positioned side by side, with one on the left and the other on the right. Above the sinks, there are two mirrors, reflecting the bathroom's interior. In addition to the sinks, there are two toothbrush holders, one on the left side and the other on the right side of the vanity. A towel is also visible, hanging on the wall above the sinks. The bathroom appears to be well-maintained and organized, providing a comfortable space for daily routines.

prompt: The image features a large, clean, and well-organized kitchen with white cabinets and a marble countertop. The kitchen is equipped with a sink, an oven, and a stove. There is a window in the kitchen, allowing natural light to enter the space. Various items can be seen throughout the kitchen, including a bowl, a vase, and a potted plant. There are also several bottles placed on the countertop, and a spoon is visible near the sink. The kitchen is well-stocked with essential items, making it a functional and inviting space.

prompt: The image features a bedroom with a large bed occupying a significant portion of the space. The bed is positioned against a wall, and there is a wooden door nearby. The room also contains a potted plant, which is placed close to the bed, adding a touch of greenery to the space. In addition to the bed and plant, there is a book on the left side of the room, possibly for reading or decoration. The room appears to be well-lit, creating a comfortable and inviting atmosphere.

## Intended uses & limitations
#### How to use
```python
# TODO: add an example code snippet for running this diffusion pipeline
```
#### Limitations and bias
[TODO: provide examples of latent issues and potential remediations]
## Training details
[TODO: describe the data used to train the model]
|
monkeyjulia0/blockassist
|
monkeyjulia0
| 2025-09-25T06:01:05Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"endangered jumping antelope",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:55:04Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- endangered jumping antelope
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
lmassaron/gemma-3-4b-finsentiment
|
lmassaron
| 2025-09-25T06:00:57Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"trl",
"sft",
"conversational",
"arxiv:1910.09700",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] |
image-text-to-text
| 2025-09-24T20:42:14Z |
---
library_name: transformers
tags:
- trl
- sft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
DidulaThavishaPro/base_test_4_sft_16bit_vllm
|
DidulaThavishaPro
| 2025-09-25T06:00:56Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-09-24T17:31:55Z |
---
base_model: unsloth/qwen2.5-coder-7b-instruct-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- qwen2
license: apache-2.0
language:
- en
---
# Uploaded finetuned model
- **Developed by:** DidulaThavishaPro
- **License:** apache-2.0
- **Finetuned from model :** unsloth/qwen2.5-coder-7b-instruct-bnb-4bit
This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
stynimshon/blockassist
|
stynimshon
| 2025-09-25T06:00:35Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"agile territorial elephant",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:14:05Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- agile territorial elephant
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
ericshanna1/blockassist
|
ericshanna1
| 2025-09-25T06:00:29Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"sneaky beaked salmon",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:53:33Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- sneaky beaked salmon
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
jfj033631/blockassist
|
jfj033631
| 2025-09-25T06:00:25Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"stubby scampering mantis",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-20T11:49:30Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- stubby scampering mantis
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Sushkrnishad/finetuning-sentiment-model-3000-samples
|
Sushkrnishad
| 2025-09-25T06:00:25Z | 0 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"text-classification",
"generated_from_trainer",
"base_model:distilbert/distilbert-base-uncased",
"base_model:finetune:distilbert/distilbert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-09-25T05:47:51Z |
---
library_name: transformers
license: apache-2.0
base_model: distilbert-base-uncased
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: finetuning-sentiment-model-3000-samples
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetuning-sentiment-model-3000-samples
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3538
- Accuracy: 0.8833
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.56.1
- Pytorch 2.8.0+cu126
- Datasets 4.0.0
- Tokenizers 0.22.0
|
liaoleonel89/blockassist
|
liaoleonel89
| 2025-09-25T05:59:57Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"hibernating nocturnal chimpanzee",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:13:33Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- hibernating nocturnal chimpanzee
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
hardyespositotw/blockassist
|
hardyespositotw
| 2025-09-25T05:59:54Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"bipedal vicious bee",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-19T08:03:55Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- bipedal vicious bee
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
vassar054/blockassist
|
vassar054
| 2025-09-25T05:59:51Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"burrowing wild clam",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:30:25Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- burrowing wild clam
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
waynekayn/blockassist
|
waynekayn
| 2025-09-25T05:59:48Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"bellowing gentle starfish",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T17:08:58Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- bellowing gentle starfish
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
DhanaR1212/finetuning-sentiment-model-3000-samples
|
DhanaR1212
| 2025-09-25T05:59:43Z | 0 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"text-classification",
"generated_from_trainer",
"base_model:distilbert/distilbert-base-uncased",
"base_model:finetune:distilbert/distilbert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-09-25T05:47:57Z |
---
library_name: transformers
license: apache-2.0
base_model: distilbert-base-uncased
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: finetuning-sentiment-model-3000-samples
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetuning-sentiment-model-3000-samples
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3500
- Accuracy: 0.86
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.56.1
- Pytorch 2.8.0+cu126
- Datasets 4.0.0
- Tokenizers 0.22.0
|
waughc080/blockassist
|
waughc080
| 2025-09-25T05:59:17Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"armored prickly caribou",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:12:52Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- armored prickly caribou
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Shubh-03/finetuning-sentiment-model-3000-samples
|
Shubh-03
| 2025-09-25T05:59:10Z | 0 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"text-classification",
"generated_from_trainer",
"base_model:distilbert/distilbert-base-uncased",
"base_model:finetune:distilbert/distilbert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-09-25T05:47:14Z |
---
library_name: transformers
license: apache-2.0
base_model: distilbert-base-uncased
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: finetuning-sentiment-model-3000-samples
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetuning-sentiment-model-3000-samples
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3371
- Accuracy: 0.8767
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.56.1
- Pytorch 2.8.0+cu126
- Datasets 4.0.0
- Tokenizers 0.22.0
|
ashleyvandemicarlos/blockassist
|
ashleyvandemicarlos
| 2025-09-25T05:59:09Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"bold keen pig",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:47:49Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- bold keen pig
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
SKIML-ICL/oursNQ-QA-QUC-Meta-Llama-3-8B-Instruct-QA-QUC-0Q-0U-0C-random-lora
|
SKIML-ICL
| 2025-09-25T05:58:34Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"generated_from_trainer",
"sft",
"trl",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"endpoints_compatible",
"region:us"
] | null | 2025-09-25T04:49:48Z |
---
base_model: meta-llama/Meta-Llama-3-8B-Instruct
library_name: transformers
model_name: oursNQ-QA-QUC-Meta-Llama-3-8B-Instruct-QA-QUC-0Q-0U-0C-random-lora
tags:
- generated_from_trainer
- sft
- trl
licence: license
---
# Model Card for oursNQ-QA-QUC-Meta-Llama-3-8B-Instruct-QA-QUC-0Q-0U-0C-random-lora
This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="SKIML-ICL/oursNQ-QA-QUC-Meta-Llama-3-8B-Instruct-QA-QUC-0Q-0U-0C-random-lora", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/Enhancing-ICL/train-LLM%7Cin-context-QA/runs/lxaqot85)
This model was trained with SFT.
### Framework versions
- TRL: 0.23.0
- Transformers: 4.56.2
- Pytorch: 2.7.1
- Datasets: 4.0.0
- Tokenizers: 0.22.1
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
varneyc352/blockassist
|
varneyc352
| 2025-09-25T05:58:05Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"melodic bipedal mongoose",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:11:43Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- melodic bipedal mongoose
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
alexshannawvnt178/blockassist
|
alexshannawvnt178
| 2025-09-25T05:57:30Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"peckish sniffing pigeon",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:44:03Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- peckish sniffing pigeon
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
ss2972583/blockassist
|
ss2972583
| 2025-09-25T05:57:05Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"grunting lively cheetah",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:27:56Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- grunting lively cheetah
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Sujith23/PPO-LunarLander-v2
|
Sujith23
| 2025-09-25T05:56:42Z | 10 | 0 |
stable-baselines3
|
[
"stable-baselines3",
"LunarLander-v2",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] |
reinforcement-learning
| 2025-09-17T13:02:41Z |
---
library_name: stable-baselines3
tags:
- LunarLander-v2
- deep-reinforcement-learning
- reinforcement-learning
- stable-baselines3
model-index:
- name: PPO
results:
- task:
type: reinforcement-learning
name: reinforcement-learning
dataset:
name: LunarLander-v2
type: LunarLander-v2
metrics:
- type: mean_reward
value: 251.92 +/- 16.45
name: mean_reward
verified: false
---
# **PPO** Agent playing **LunarLander-v2**
This is a trained model of a **PPO** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
|
mazesmazes/tiny-audio
|
mazesmazes
| 2025-09-25T05:56:37Z | 36 | 0 | null |
[
"safetensors",
"asr_model",
"automatic-speech-recognition",
"custom_code",
"en",
"dataset:mozilla-foundation/common_voice_17_0",
"dataset:speechcolab/gigaspeech",
"dataset:openslr/librispeech_asr",
"base_model:HuggingFaceTB/SmolLM2-360M-Instruct",
"base_model:finetune:HuggingFaceTB/SmolLM2-360M-Instruct",
"license:mit",
"region:us"
] |
automatic-speech-recognition
| 2025-09-23T16:27:59Z |
---
license: mit
datasets:
- mozilla-foundation/common_voice_17_0
- speechcolab/gigaspeech
- openslr/librispeech_asr
language:
- en
metrics:
- wer
base_model:
- openai/whisper-small
- HuggingFaceTB/SmolLM2-360M-Instruct
pipeline_tag: automatic-speech-recognition
---
# Tiny Audio - Whisper-SmolLM2 ASR Model
A lightweight ASR model combining Whisper-small encoder with SmolLM2 decoder, trained with LoRA for parameter-efficient fine-tuning (~300 lines of code).
## Quick Start
```python
from transformers import AutoModelForSpeechSeq2Seq, pipeline
# Load model
model = AutoModelForSpeechSeq2Seq.from_pretrained(
"mazesmazes/tiny-audio",
trust_remote_code=True
)
# Create pipeline and transcribe
asr = pipeline("automatic-speech-recognition", model=model)
result = asr("audio.wav")
print(result["text"])
```
## Architecture
- **Encoder**: Frozen Whisper-small (39M params)
- **Decoder**: SmolLM2-360M with LoRA adapters (7.3M trainable params)
- **Total**: ~400M parameters, only 2% trained
## Training
- **Data**: LibriSpeech, GigaSpeech, Common Voice (English)
- **Method**: LoRA rank 32-64, BF16 mixed precision
- **Performance**: ~15-20% WER on LibriSpeech test-clean
## Links
- **Code**: [github.com/alexkroman/tiny-audio](https://github.com/alexkroman/tiny-audio)
- **Demo**: `demo/app.py` in repository
- **License**: MIT
|
jfdh0843/blockassist
|
jfdh0843
| 2025-09-25T05:56:18Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"elusive hardy antelope",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-20T11:42:46Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- elusive hardy antelope
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF
|
mradermacher
| 2025-09-25T05:56:12Z | 0 | 0 |
transformers
|
[
"transformers",
"gguf",
"programming",
"code generation",
"code",
"coding",
"coder",
"chat",
"brainstorm",
"qwen",
"qwen3",
"qwencoder",
"brainstorm 20x",
"creative",
"all uses cases",
"Jan-V1",
"horror",
"science fiction",
"fantasy",
"Star Trek",
"Star Trek Original",
"Star Trek The Next Generation",
"Star Trek Deep Space Nine",
"Star Trek Voyager",
"Star Trek Enterprise",
"Star Trek Discovery.",
"finetune",
"thinking",
"reasoning",
"unsloth",
"6x6B",
"moe",
"mixture of experts",
"en",
"dataset:DavidAU/horror-nightmare1",
"dataset:DavidAU/ST-Org",
"dataset:DavidAU/ST-TNG",
"dataset:DavidAU/ST-DS9",
"dataset:DavidAU/ST-VOY",
"dataset:DavidAU/ST-ENT",
"dataset:DavidAU/ST-DIS",
"base_model:DavidAU/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B",
"base_model:quantized:DavidAU/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B",
"license:apache-2.0",
"endpoints_compatible",
"region:us",
"conversational"
] | null | 2025-09-24T11:31:12Z |
---
base_model: DavidAU/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B
datasets:
- DavidAU/horror-nightmare1
- DavidAU/ST-Org
- DavidAU/ST-TNG
- DavidAU/ST-DS9
- DavidAU/ST-VOY
- DavidAU/ST-ENT
- DavidAU/ST-DIS
language:
- en
library_name: transformers
license: apache-2.0
mradermacher:
readme_rev: 1
quantized_by: mradermacher
tags:
- programming
- code generation
- code
- coding
- coder
- chat
- code
- chat
- brainstorm
- qwen
- qwen3
- qwencoder
- brainstorm 20x
- creative
- all uses cases
- Jan-V1
- horror
- science fiction
- fantasy
- Star Trek
- Star Trek Original
- Star Trek The Next Generation
- Star Trek Deep Space Nine
- Star Trek Voyager
- Star Trek Enterprise
- Star Trek Discovery.
- finetune
- thinking
- reasoning
- unsloth
- 6x6B
- moe
- mixture of experts
- finetune
---
## About
<!-- ### quantize_version: 2 -->
<!-- ### output_tensor_quantised: 1 -->
<!-- ### convert_type: hf -->
<!-- ### vocab_type: -->
<!-- ### tags: -->
<!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
<!-- ### quants_skip: -->
<!-- ### skip_mmproj: -->
static quants of https://huggingface.co/DavidAU/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B
<!-- provided-files -->
***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF).***
weighted/imatrix quants are available at https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-i1-GGUF
## Usage
If you are unsure how to use GGUF files, refer to one of [TheBloke's
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
more details, including on how to concatenate multi-part files.
## Provided Quants
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
| Link | Type | Size/GB | Notes |
|:-----|:-----|--------:|:------|
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q2_K.gguf) | Q2_K | 10.0 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q3_K_S.gguf) | Q3_K_S | 11.8 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q3_K_M.gguf) | Q3_K_M | 13.1 | lower quality |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q3_K_L.gguf) | Q3_K_L | 14.1 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.IQ4_XS.gguf) | IQ4_XS | 14.7 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q4_K_S.gguf) | Q4_K_S | 15.4 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q4_K_M.gguf) | Q4_K_M | 16.4 | fast, recommended |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q5_K_S.gguf) | Q5_K_S | 18.6 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q5_K_M.gguf) | Q5_K_M | 19.2 | |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q6_K.gguf) | Q6_K | 22.2 | very good quality |
| [GGUF](https://huggingface.co/mradermacher/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B-GGUF/resolve/main/Qwen3-MOE-6x6B-Star-Trek-Universe-Alpha-D-256k-ctx-36B.Q8_0.gguf) | Q8_0 | 28.7 | fast, best quality |
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):

And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
## FAQ / Model Request
See https://huggingface.co/mradermacher/model_requests for some answers to
questions you might have and/or if you want some other model quantized.
## Thanks
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
me use its servers and providing upgrades to my workstation to enable
this work in my free time.
<!-- end -->
|
xyzasdfghjkl123456/VideoJudge-3B
|
xyzasdfghjkl123456
| 2025-09-25T05:55:59Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2_5_vl",
"image-to-text",
"arxiv:1910.09700",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
image-to-text
| 2025-09-25T05:55:12Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
tliomh002/blockassist
|
tliomh002
| 2025-09-25T05:55:37Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"bold mimic raccoon",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:09:57Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- bold mimic raccoon
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
DidulaThavishaPro/base_test_4_sft_8bit_vllm
|
DidulaThavishaPro
| 2025-09-25T05:55:29Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"qwen2",
"en",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2025-09-24T17:21:24Z |
---
base_model: unsloth/qwen2.5-coder-7b-instruct-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- qwen2
license: apache-2.0
language:
- en
---
# Uploaded finetuned model
- **Developed by:** DidulaThavishaPro
- **License:** apache-2.0
- **Finetuned from model :** unsloth/qwen2.5-coder-7b-instruct-bnb-4bit
This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
milindk0404/finetuning-sentiment-model-3000-samples
|
milindk0404
| 2025-09-25T05:55:13Z | 0 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"text-classification",
"generated_from_trainer",
"base_model:distilbert/distilbert-base-uncased",
"base_model:finetune:distilbert/distilbert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-09-25T05:42:53Z |
---
library_name: transformers
license: apache-2.0
base_model: distilbert-base-uncased
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: finetuning-sentiment-model-3000-samples
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# finetuning-sentiment-model-3000-samples
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4548
- Accuracy: 0.8633
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
### Framework versions
- Transformers 4.56.1
- Pytorch 2.8.0+cu126
- Datasets 4.0.0
- Tokenizers 0.22.0
|
dgvd28533/blockassist
|
dgvd28533
| 2025-09-25T05:54:48Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"freckled monstrous ape",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-20T11:33:41Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- freckled monstrous ape
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
doomzillwillie/blockassist
|
doomzillwillie
| 2025-09-25T05:54:23Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"foxy twitchy horse",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:31:25Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- foxy twitchy horse
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
singaron34/blockassist
|
singaron34
| 2025-09-25T05:54:11Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"fast sleek seahorse",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:24:58Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- fast sleek seahorse
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
janealexisbrandy/blockassist
|
janealexisbrandy
| 2025-09-25T05:53:39Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"screeching camouflaged lobster",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:30:49Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- screeching camouflaged lobster
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
corzamennav/blockassist-bc-territorial_wild_antelope_1758779538
|
corzamennav
| 2025-09-25T05:53:26Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"territorial wild antelope",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-25T05:53:20Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- territorial wild antelope
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
baldo3615/blockassist
|
baldo3615
| 2025-09-25T05:52:40Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"fierce armored albatross",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:07:24Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- fierce armored albatross
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
rinm74863/blockassist
|
rinm74863
| 2025-09-25T05:52:29Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"frisky placid seal",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:23:42Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- frisky placid seal
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
brockettye/blockassist
|
brockettye
| 2025-09-25T05:51:57Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"dextrous running goose",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:06:53Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- dextrous running goose
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
crsunnips86/blockassist
|
crsunnips86
| 2025-09-25T05:51:49Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"short vocal ant",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:23:08Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- short vocal ant
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
scottodomingo85/blockassist
|
scottodomingo85
| 2025-09-25T05:51:19Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"rabid lively mongoose",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:06:14Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- rabid lively mongoose
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
runaligsa463/blockassist
|
runaligsa463
| 2025-09-25T05:50:39Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"prowling regal bat",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:22:13Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- prowling regal bat
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Enkhmanlai/mmbert-xnli
|
Enkhmanlai
| 2025-09-25T05:50:37Z | 0 | 0 |
transformers
|
[
"transformers",
"tensorboard",
"safetensors",
"modernbert",
"text-classification",
"generated_from_trainer",
"base_model:jhu-clsp/mmBERT-small",
"base_model:finetune:jhu-clsp/mmBERT-small",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-classification
| 2025-09-25T05:50:29Z |
---
library_name: transformers
license: mit
base_model: jhu-clsp/mmBERT-small
tags:
- generated_from_trainer
metrics:
- accuracy
- f1
model-index:
- name: mmbert-xnli
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mmbert-xnli
This model is a fine-tuned version of [jhu-clsp/mmBERT-small](https://huggingface.co/jhu-clsp/mmBERT-small) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3736
- Accuracy: 0.8863
- F1: 0.8863
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| No log | 1.0 | 337 | 0.4502 | 0.8600 | 0.8587 |
| 0.6169 | 2.0 | 674 | 0.3983 | 0.8796 | 0.8786 |
| 0.2926 | 3.0 | 1011 | 0.3693 | 0.8904 | 0.8907 |
### Framework versions
- Transformers 4.52.4
- Pytorch 2.6.0+cu124
- Datasets 3.6.0
- Tokenizers 0.21.2
|
zacherysix660/blockassist
|
zacherysix660
| 2025-09-25T05:50:35Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"toothy gilded gorilla",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:05:40Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- toothy gilded gorilla
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
maribethalysenshontaitorrell/blockassist
|
maribethalysenshontaitorrell
| 2025-09-25T05:50:29Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"leggy bipedal lizard",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:18:55Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- leggy bipedal lizard
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
xyzasdfghjkl123456/VideoJudgeR-3B
|
xyzasdfghjkl123456
| 2025-09-25T05:50:15Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2_5_vl",
"image-to-text",
"arxiv:1910.09700",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
image-to-text
| 2025-09-25T05:49:22Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
bjdsau/blockassist
|
bjdsau
| 2025-09-25T05:50:02Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"stinky lumbering butterfly",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:21:40Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- stinky lumbering butterfly
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
ywoneonj201/blockassist
|
ywoneonj201
| 2025-09-25T05:49:58Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"soft grassy penguin",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:05:01Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- soft grassy penguin
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
cameroneltamadoerikqxc75/blockassist
|
cameroneltamadoerikqxc75
| 2025-09-25T05:49:53Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"soft giant macaw",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:18:19Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- soft giant macaw
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
bibb56334/blockassist
|
bibb56334
| 2025-09-25T05:49:24Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"vigilant yapping fly",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:04:27Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- vigilant yapping fly
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
jikerrums/blockassist
|
jikerrums
| 2025-09-25T05:49:13Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"bold humming crab",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:21:10Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- bold humming crab
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
inferencerlabs/deepseek-v3.1-Terminus-MLX-5.5bit
|
inferencerlabs
| 2025-09-25T05:48:41Z | 150 | 0 |
mlx
|
[
"mlx",
"safetensors",
"deepseek_v3",
"text-generation",
"conversational",
"custom_code",
"base_model:deepseek-ai/DeepSeek-V3.1-Terminus",
"base_model:quantized:deepseek-ai/DeepSeek-V3.1-Terminus",
"license:mit",
"5-bit",
"region:us"
] |
text-generation
| 2025-09-23T09:12:47Z |
---
license: mit
library_name: mlx
base_model: deepseek-ai/DeepSeek-V3.1-Terminus
tags:
- mlx
pipeline_tag: text-generation
---
** CURRENTLY UPLOADING **
**See DeepSeek-V3.1-Terminus 5.5bit MLX in action - [demonstration video coming soon](https://youtube.com/xcreate)**
*q5.5bit quant typically achieves 1.141 perplexity in our testing*
| Quantization | Perplexity |
|:------------:|:----------:|
| **q2.5** | 41.293 |
| **q3.5** | 1.900 |
| **q4.5** | 1.168 |
| **q5.5** | 1.141 |
| **q6.5** | 1.128 |
| **q8.5** | 1.128 |
## Usage Notes
* [Original DeepSeek V3.1](https://huggingface.co/inferencerlabs/deepseek-v3.1-MLX-5.5bit) performed better in our testing
* Runs on a single M3 Ultra 512GB RAM using [Inferencer app](https://inferencer.com)
* Memory usage: ~480 GB
* Expect ~13-19 tokens/s
* Quantized with a modified version of [MLX](https://github.com/ml-explore/mlx) 0.27
* For more details see [demonstration video coming soon](https://youtube.com/xcreate) or visit [DeepSeek-V3.1-Terminus](https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Terminus).
## Disclaimer
We are not the creator, originator, or owner of any model listed. Each model is created and provided by third parties. Models may not always be accurate or contextually appropriate. You are responsible for verifying the information before making important decisions. We are not liable for any damages, losses, or issues arising from its use, including data loss or inaccuracies in AI-generated content.
|
ilorajahan8/blockassist
|
ilorajahan8
| 2025-09-25T05:48:25Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"feathered exotic boar",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-20T11:56:50Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- feathered exotic boar
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
moosclemente34/blockassist
|
moosclemente34
| 2025-09-25T05:48:08Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"pawing grunting toucan",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:13:33Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- pawing grunting toucan
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
nt250170/blockassist
|
nt250170
| 2025-09-25T05:47:47Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"agile moist baboon",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:19:46Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- agile moist baboon
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
sidhantoon/Koto6
|
sidhantoon
| 2025-09-25T05:47:47Z | 0 | 0 | null |
[
"safetensors",
"any-to-any",
"omega",
"omegalabs",
"bittensor",
"agi",
"license:mit",
"region:us"
] |
any-to-any
| 2025-09-25T05:41:23Z |
---
license: mit
tags:
- any-to-any
- omega
- omegalabs
- bittensor
- agi
---
This is an Any-to-Any model checkpoint for the OMEGA Labs x Bittensor Any-to-Any subnet.
Check out the [git repo](https://github.com/omegalabsinc/omegalabs-anytoany-bittensor) and find OMEGA on X: [@omegalabsai](https://x.com/omegalabsai).
|
tonikeuoj/blockassist
|
tonikeuoj
| 2025-09-25T05:47:29Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"long graceful dragonfly",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:02:37Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- long graceful dragonfly
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
meyintson3/blockassist
|
meyintson3
| 2025-09-25T05:47:25Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"pensive lively armadillo",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:12:53Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- pensive lively armadillo
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
folksz218/blockassist
|
folksz218
| 2025-09-25T05:46:52Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"dextrous barky chicken",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T15:11:49Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- dextrous barky chicken
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
sidhantoon/Koto5
|
sidhantoon
| 2025-09-25T05:46:49Z | 0 | 0 | null |
[
"safetensors",
"any-to-any",
"omega",
"omegalabs",
"bittensor",
"agi",
"license:mit",
"region:us"
] |
any-to-any
| 2025-09-25T05:41:17Z |
---
license: mit
tags:
- any-to-any
- omega
- omegalabs
- bittensor
- agi
---
This is an Any-to-Any model checkpoint for the OMEGA Labs x Bittensor Any-to-Any subnet.
Check out the [git repo](https://github.com/omegalabsinc/omegalabs-anytoany-bittensor) and find OMEGA on X: [@omegalabsai](https://x.com/omegalabsai).
|
johngreendr1/d12f3bc9-5592-4f60-967b-fc98d4ca0ba1
|
johngreendr1
| 2025-09-25T05:46:44Z | 0 | 0 |
peft
|
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:unsloth/gemma-2-2b-it",
"base_model:adapter:unsloth/gemma-2-2b-it",
"region:us"
] | null | 2025-09-25T04:26:44Z |
---
base_model: unsloth/gemma-2-2b-it
library_name: peft
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.15.1
|
opete638/blockassist
|
opete638
| 2025-09-25T05:46:43Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"rough noisy mouse",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:02:00Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- rough noisy mouse
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
alinearmun/blockassist
|
alinearmun
| 2025-09-25T05:46:28Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"long rugged ant",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:18:47Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- long rugged ant
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
thu136093/blockassist-bc-waddling_cunning_slug_1758778301
|
thu136093
| 2025-09-25T05:45:15Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"waddling cunning slug",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-25T05:45:10Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- waddling cunning slug
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
carrollg679/blockassist
|
carrollg679
| 2025-09-25T05:44:51Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"sizable cunning narwhal",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T04:00:19Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- sizable cunning narwhal
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
AlexanderArtT/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-tiny_nimble_warthog
|
AlexanderArtT
| 2025-09-25T05:44:29Z | 14 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"rl-swarm",
"grpo",
"gensyn",
"I am tiny nimble warthog",
"trl",
"genrl-swarm",
"I am tiny_nimble_warthog",
"conversational",
"arxiv:2402.03300",
"base_model:unsloth/Qwen2.5-0.5B-Instruct",
"base_model:finetune:unsloth/Qwen2.5-0.5B-Instruct",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-05-13T22:11:38Z |
---
base_model: unsloth/Qwen2.5-0.5B-Instruct
library_name: transformers
model_name: Qwen2.5-0.5B-Instruct-Gensyn-Swarm-tiny_nimble_warthog
tags:
- generated_from_trainer
- rl-swarm
- grpo
- gensyn
- I am tiny nimble warthog
- trl
- genrl-swarm
- I am tiny_nimble_warthog
licence: license
---
# Model Card for Qwen2.5-0.5B-Instruct-Gensyn-Swarm-tiny_nimble_warthog
This model is a fine-tuned version of [unsloth/Qwen2.5-0.5B-Instruct](https://huggingface.co/unsloth/Qwen2.5-0.5B-Instruct).
It has been trained using [TRL](https://github.com/huggingface/trl).
## Quick start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="AlexanderArtT/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-tiny_nimble_warthog", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training procedure
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
### Framework versions
- TRL: 0.17.0
- Transformers: 4.51.3
- Pytorch: 2.7.0
- Datasets: 3.6.0
- Tokenizers: 0.21.1
## Citations
Cite GRPO as:
```bibtex
@article{zhihong2024deepseekmath,
title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
year = 2024,
eprint = {arXiv:2402.03300},
}
```
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
```
|
rupatomes8/blockassist
|
rupatomes8
| 2025-09-25T05:44:23Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"eager loud porcupine",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:17:08Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- eager loud porcupine
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
ottleyemory/blockassist
|
ottleyemory
| 2025-09-25T05:43:47Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"hoarse lumbering sparrow",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-19T07:56:26Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- hoarse lumbering sparrow
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
terrancekohler969/blockassist
|
terrancekohler969
| 2025-09-25T05:43:36Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"waddling lethal prawn",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:59:11Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- waddling lethal prawn
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
corzamennav/blockassist-bc-territorial_wild_antelope_1758778922
|
corzamennav
| 2025-09-25T05:43:11Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"territorial wild antelope",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-25T05:43:05Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- territorial wild antelope
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Novaciano/Qwen3-VL-1B-Merged-Q5_K_M-GGUF
|
Novaciano
| 2025-09-25T05:42:50Z | 0 | 0 |
transformers
|
[
"transformers",
"gguf",
"multimodal",
"vision-language",
"qwen3",
"qwen2.5-vl",
"llama-cpp",
"gguf-my-repo",
"image-text-to-text",
"en",
"base_model:ViFortune-AI/Qwen3-VL-1B-Merged",
"base_model:quantized:ViFortune-AI/Qwen3-VL-1B-Merged",
"endpoints_compatible",
"region:us",
"conversational"
] |
image-text-to-text
| 2025-09-25T05:42:43Z |
---
license_name: vifortune-research
license_link: https://huggingface.co/Qwen/Qwen3-VL-1B/blob/main/LICENSE
language:
- en
pipeline_tag: image-text-to-text
tags:
- multimodal
- vision-language
- qwen3
- qwen2.5-vl
- llama-cpp
- gguf-my-repo
library_name: transformers
base_model: ViFortune-AI/Qwen3-VL-1B-Merged
---
# Novaciano/Qwen3-VL-1B-Merged-Q5_K_M-GGUF
This model was converted to GGUF format from [`ViFortune-AI/Qwen3-VL-1B-Merged`](https://huggingface.co/ViFortune-AI/Qwen3-VL-1B-Merged) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
Refer to the [original model card](https://huggingface.co/ViFortune-AI/Qwen3-VL-1B-Merged) for more details on the model.
## Use with llama.cpp
Install llama.cpp through brew (works on Mac and Linux)
```bash
brew install llama.cpp
```
Invoke the llama.cpp server or the CLI.
### CLI:
```bash
llama-cli --hf-repo Novaciano/Qwen3-VL-1B-Merged-Q5_K_M-GGUF --hf-file qwen3-vl-1b-merged-q5_k_m.gguf -p "The meaning to life and the universe is"
```
### Server:
```bash
llama-server --hf-repo Novaciano/Qwen3-VL-1B-Merged-Q5_K_M-GGUF --hf-file qwen3-vl-1b-merged-q5_k_m.gguf -c 2048
```
Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
Step 1: Clone llama.cpp from GitHub.
```
git clone https://github.com/ggerganov/llama.cpp
```
Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
```
cd llama.cpp && LLAMA_CURL=1 make
```
Step 3: Run inference through the main binary.
```
./llama-cli --hf-repo Novaciano/Qwen3-VL-1B-Merged-Q5_K_M-GGUF --hf-file qwen3-vl-1b-merged-q5_k_m.gguf -p "The meaning to life and the universe is"
```
or
```
./llama-server --hf-repo Novaciano/Qwen3-VL-1B-Merged-Q5_K_M-GGUF --hf-file qwen3-vl-1b-merged-q5_k_m.gguf -c 2048
```
|
sandovalgregg08/blockassist
|
sandovalgregg08
| 2025-09-25T05:42:25Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"sprightly twitchy donkey",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:57:49Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- sprightly twitchy donkey
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
samanta-scratch/wav2vec2-large-mms-1b-dv
|
samanta-scratch
| 2025-09-25T05:42:19Z | 0 | 0 |
transformers
|
[
"transformers",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-09-25T05:42:18Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
WelcomeDarrell399/blockassist
|
WelcomeDarrell399
| 2025-09-25T05:42:11Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"stocky pawing bear",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T16:52:50Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- stocky pawing bear
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Novaciano/Qwen3-VL-1B-Merged-Q8_0-GGUF
|
Novaciano
| 2025-09-25T05:42:11Z | 0 | 0 |
transformers
|
[
"transformers",
"gguf",
"multimodal",
"vision-language",
"qwen3",
"qwen2.5-vl",
"llama-cpp",
"gguf-my-repo",
"image-text-to-text",
"en",
"base_model:ViFortune-AI/Qwen3-VL-1B-Merged",
"base_model:quantized:ViFortune-AI/Qwen3-VL-1B-Merged",
"endpoints_compatible",
"region:us",
"conversational"
] |
image-text-to-text
| 2025-09-25T05:42:04Z |
---
license_name: vifortune-research
license_link: https://huggingface.co/Qwen/Qwen3-VL-1B/blob/main/LICENSE
language:
- en
pipeline_tag: image-text-to-text
tags:
- multimodal
- vision-language
- qwen3
- qwen2.5-vl
- llama-cpp
- gguf-my-repo
library_name: transformers
base_model: ViFortune-AI/Qwen3-VL-1B-Merged
---
# Novaciano/Qwen3-VL-1B-Merged-Q8_0-GGUF
This model was converted to GGUF format from [`ViFortune-AI/Qwen3-VL-1B-Merged`](https://huggingface.co/ViFortune-AI/Qwen3-VL-1B-Merged) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
Refer to the [original model card](https://huggingface.co/ViFortune-AI/Qwen3-VL-1B-Merged) for more details on the model.
## Use with llama.cpp
Install llama.cpp through brew (works on Mac and Linux)
```bash
brew install llama.cpp
```
Invoke the llama.cpp server or the CLI.
### CLI:
```bash
llama-cli --hf-repo Novaciano/Qwen3-VL-1B-Merged-Q8_0-GGUF --hf-file qwen3-vl-1b-merged-q8_0.gguf -p "The meaning to life and the universe is"
```
### Server:
```bash
llama-server --hf-repo Novaciano/Qwen3-VL-1B-Merged-Q8_0-GGUF --hf-file qwen3-vl-1b-merged-q8_0.gguf -c 2048
```
Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
Step 1: Clone llama.cpp from GitHub.
```
git clone https://github.com/ggerganov/llama.cpp
```
Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
```
cd llama.cpp && LLAMA_CURL=1 make
```
Step 3: Run inference through the main binary.
```
./llama-cli --hf-repo Novaciano/Qwen3-VL-1B-Merged-Q8_0-GGUF --hf-file qwen3-vl-1b-merged-q8_0.gguf -p "The meaning to life and the universe is"
```
or
```
./llama-server --hf-repo Novaciano/Qwen3-VL-1B-Merged-Q8_0-GGUF --hf-file qwen3-vl-1b-merged-q8_0.gguf -c 2048
```
|
gdg98574/blockassist
|
gdg98574
| 2025-09-25T05:42:10Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"lazy zealous dolphin",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-20T11:30:37Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- lazy zealous dolphin
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
hzyitong45/blockassist
|
hzyitong45
| 2025-09-25T05:41:42Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"armored climbing seahorse",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T16:52:25Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- armored climbing seahorse
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
do0090022/blockassist
|
do0090022
| 2025-09-25T05:41:12Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"sharp voracious sealion",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:56:33Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- sharp voracious sealion
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Jax-dan/SmolVLM-0.6B-Cap
|
Jax-dan
| 2025-09-25T05:41:04Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"image-text-to-text",
"zh",
"en",
"dataset:lmms-lab/COCO-Caption2017",
"base_model:HuggingFaceTB/SmolVLM-256M-Instruct",
"base_model:finetune:HuggingFaceTB/SmolVLM-256M-Instruct",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
image-text-to-text
| 2025-09-25T05:32:15Z |
---
license: apache-2.0
datasets:
- lmms-lab/COCO-Caption2017
language:
- zh
- en
metrics:
- bertscore
- bleu
- rouge
- meteor
base_model:
- HuggingFaceTB/SmolVLM-256M-Instruct
- Qwen/Qwen2.5-0.5B-Instruct
pipeline_tag: image-text-to-text
library_name: transformers
---
|
hzyitong44/blockassist
|
hzyitong44
| 2025-09-25T05:41:02Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"secretive scavenging rhino",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T16:51:47Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- secretive scavenging rhino
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
hzyitong43/blockassist
|
hzyitong43
| 2025-09-25T05:40:27Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"mighty polished armadillo",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T16:51:11Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- mighty polished armadillo
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
feltonv60/blockassist
|
feltonv60
| 2025-09-25T05:40:26Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"moist knobby octopus",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:55:50Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- moist knobby octopus
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
fredericlaclair19/blockassist
|
fredericlaclair19
| 2025-09-25T05:39:51Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"feline jagged antelope",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:55:16Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- feline jagged antelope
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
hzyitong35/blockassist
|
hzyitong35
| 2025-09-25T05:39:43Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"amphibious burrowing okapi",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T16:50:32Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- amphibious burrowing okapi
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
bunrsmiker21/blockassist
|
bunrsmiker21
| 2025-09-25T05:39:29Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"omnivorous lightfooted gerbil",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:13:35Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- omnivorous lightfooted gerbil
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
ganga4364/whisper-small-tibetan-wylie-checkpoint-4000
|
ganga4364
| 2025-09-25T05:39:18Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"whisper",
"automatic-speech-recognition",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] |
automatic-speech-recognition
| 2025-09-25T05:39:01Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
hnewcomb637/blockassist
|
hnewcomb637
| 2025-09-25T05:39:13Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"mute flexible caterpillar",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:54:33Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- mute flexible caterpillar
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
keithmansell71/blockassist
|
keithmansell71
| 2025-09-25T05:38:39Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"aquatic shy elephant",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:53:50Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- aquatic shy elephant
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
funesnisren366/blockassist
|
funesnisren366
| 2025-09-25T05:38:38Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"beaked bristly nightingale",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:13:07Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- beaked bristly nightingale
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
runspumes/blockassist
|
runspumes
| 2025-09-25T05:38:00Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"stalking savage spider",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:12:35Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- stalking savage spider
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
lsalaam568/blockassist
|
lsalaam568
| 2025-09-25T05:37:59Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"waddling feathered mallard",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:53:16Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- waddling feathered mallard
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
haihp02/8182b079-7371-4aba-a126-4f6af60546cf
|
haihp02
| 2025-09-25T05:37:58Z | 0 | 0 |
transformers
|
[
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2025-09-25T03:24:02Z |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
hzyitong32/blockassist
|
hzyitong32
| 2025-09-25T05:37:41Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"skittish regal otter",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-23T16:48:31Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- skittish regal otter
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
xnvl/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-rugged_slithering_termite
|
xnvl
| 2025-09-25T05:37:35Z | 10 | 0 |
transformers
|
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"rl-swarm",
"genrl-swarm",
"grpo",
"gensyn",
"I am rugged_slithering_termite",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-09-18T04:55:30Z |
---
library_name: transformers
tags:
- rl-swarm
- genrl-swarm
- grpo
- gensyn
- I am rugged_slithering_termite
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
|
H5N1AIDS/Transcribe_and_Translate_Subtitles
|
H5N1AIDS
| 2025-09-25T05:37:17Z | 1 | 3 | null |
[
"onnx",
"region:us"
] | null | 2025-05-08T03:51:45Z |
## 🎬 视频字幕转录和翻译 / Transcribe and Translate Subtitles
**一个强大的、隐私优先的视频字幕转录和翻译工具**
</br>
**A powerful, privacy-first tool for transcribing and translating video subtitles**
[](https://github.com/your-repo)
[](https://onnxruntime.ai/)
[](https://github.com/your-repo)
[Visit Github](https://github.com/DakeQQ/Transcribe-and-Translate-Subtitles)
---
## 🔒 隐私保证 / Privacy Guarantee
> **🚨 所有处理完全离线运行 / All processing runs completely offline**<br>
> - 无需互联网连接,确保最大程度的隐私和数据安全<br>
> - No internet connection required, ensuring maximum privacy and data security.
---
## 🚀 快速入门 / Quick Start
### 环境准备 / Prerequisites
```bash
# 安装 FFmpeg / Install FFmpeg
conda install ffmpeg
pip install -r requirements.txt
# 安装 Python 依赖 / Install Python dependencies
# 请根据您的硬件平台安装正确的包 / Please according to your hardware platform install the right package
# ----------------------------------------
# For CPU only
# onnxruntime>=1.22.0
# ----------------------------------------
# For Linux + AMD
# 请先按照 URL 设置 ROCm / Please follow the URL to set up the ROCm first before pip install onnxruntime-rocm
# https://rocm.docs.amd.com/projects/radeon/en/latest/docs/install/native_linux/install-onnx.html
# https://onnxruntime.ai/docs/execution-providers/Vitis-AI-ExecutionProvider.html
# onnxruntime>=1.22.0
# onnxruntime-rocm>=1.22.0
# ----------------------------------------
# For Windows + (Intel or AMD)
# onnxruntime>=1.22.0
# onnxruntime-directml>=1.22.0
# ----------------------------------------
# For Intel OpenVINO CPU & GPU & NPU
# onnxruntime>=1.22.0
# onnxruntime-openvino>=1.22.0
# ----------------------------------------
# For NVIDIA-CUDA
# onnxruntime>=1.22.0
# onnxruntime-gpu>=1.22.0
# ----------------------------------------
```
### 设置
1. **下载模型**: 从 [HuggingFace](https://huggingface.co/H5N1AIDS/Transcribe_and_Translate_Subtitles) 获取所需模型 ,只下载您想要的模型并保持文件夹路径与当前定义相同,无需全部下载。
2. **下载脚本**: 将 `run.py` 放置在您的 `Transcribe_and_Translate_Subtitles` 文件夹中
3. **添加媒体**: 将您的音视频放置在 `Transcribe_and_Translate_Subtitles/Media/` 目录下
4. **运行**: 务必在`Transcribe_and_Translate_Subtitles`目录下执行 `python run.py` 并打开 Web 界面
### Setup
1. **Download Models**: Get the required models from [HuggingFace](https://huggingface.co/H5N1AIDS/Transcribe_and_Translate_Subtitles). Download only the models you want and keep the folder path the same as currently defined, no need to download all.
2. **Download Script**: Place `run.py` in your `Transcribe_and_Translate_Subtitles` folder
3. **Add Media**: Place your audios/videos in `Transcribe_and_Translate_Subtitles/Media/`
4. **Run**: You must execute `python run.py` in the `Transcribe_and_Translate_Subtitles` folder and open the web interface
5.
### 结果 / Results
在以下位置找到您处理后的字幕 / Find your processed subtitles in:
```
Transcribe_and_Translate_Subtitles/Results/Subtitles/
```
**准备好开始了吗?/ Ready to get started?** 🎉
---
## ✨ 功能特性 / Features
### 🔇 降噪模型 / Noise Reduction Models
- **[DFSMN](https://modelscope.cn/models/iic/speech_dfsmn_ans_psm_48k_causal)**
- **[GTCRN](https://github.com/Xiaobin-Rong/gtcrn)**
- **[ZipEnhancer](https://modelscope.cn/models/iic/speech_zipenhancer_ans_multiloss_16k_base)**
- **[Mel-Band-Roformer](https://github.com/KimberleyJensen/Mel-Band-Roformer-Vocal-Model)**
- **[MossFormerGAN_SE_16K](https://www.modelscope.cn/models/alibabasglab/MossFormerGAN_SE_16K)**
- **[MossFormer2_SE_48K](https://www.modelscope.cn/models/alibabasglab/MossFormer2_SE_48K)**
### 🎤 语音活动检测 (VAD) / Voice Activity Detection (VAD)
- **[Faster-Whisper-Silero](https://github.com/SYSTRAN/faster-whisper/blob/master/faster_whisper/vad.py)**
- **[Official-Silero-v6](https://github.com/snakers4/silero-vad)**
- **[HumAware](https://huggingface.co/CuriousMonkey7/HumAware-VAD)**
- **[NVIDIA-NeMo-VAD-v2.0](https://huggingface.co/nvidia/Frame_VAD_Multilingual_MarbleNet_v2.0)**
- **[TEN-VAD](https://github.com/TEN-framework/ten-vad)**
- **[Pyannote-Segmentation-3.0](https://huggingface.co/pyannote/segmentation-3.0)**
- *注意:您需要接受 Pyannote 的使用条款并下载 Pyannote 的 `pytorch_model.bin` 文件。将其放置在 `VAD/pyannote_segmentation` 文件夹中*。
- *Note: You need to accept Pyannote's terms of use and download the Pyannote `pytorch_model.bin` file. Place it in the `VAD/pyannote_segmentation` folder.*
### 🗣️ 语音识别 (ASR) / Speech Recognition (ASR)
#### 多语言模型 / Multilingual Models
- **[SenseVoice-Small-Multilingual](https://modelscope.cn/models/iic/SenseVoiceSmall)**
- **[Dolphin-Small-Asian 亚洲语言](https://github.com/DataoceanAI/Dolphin)**
- **[Paraformer-Large-Chinese 中文](https://modelscope.cn/models/iic/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch)**
- **[Paraformer-Large-English 英语](https://modelscope.cn/models/iic/speech_paraformer_asr-en-16k-vocab4199-pytorch)**
- **[FireRedASR-AED-L Chinese 中文](https://github.com/FireRedTeam/FireRedASR)**
- **[Official-Whisper-Large-v3-Multilingual](https://huggingface.co/openai/whisper-large-v3)**
- **[Official-Whisper-Large-v3-Turbo-Multilingual](https://huggingface.co/openai/whisper-large-v3-turbo)**
- **[阿拉伯语 / Arabic](https://huggingface.co/Byne/whisper-large-v3-arabic)**
- **[巴斯克语 / Basque](https://huggingface.co/xezpeleta/whisper-large-v3-eu)**
- **[粤语 / Cantonese-Yue](https://huggingface.co/JackyHoCL/whisper-large-v3-turbo-cantonese-yue-english)**
- **[中文 / Chinese](https://huggingface.co/BELLE-2/Belle-whisper-large-v3-zh-punct)**
- **[台湾客家话 / Chinese-Hakka](https://huggingface.co/formospeech/whisper-large-v3-taiwanese-hakka)**
- **[台湾闽南语 / Chinese-Minnan](https://huggingface.co/TSukiLen/whisper-medium-chinese-tw-minnan)**
- **[台湾华语 / Chinese-Taiwan](https://huggingface.co/JacobLinCool/whisper-large-v3-turbo-common_voice_19_0-zh-TW)**
- **[CrisperWhisper-Multilingual](https://github.com/nyrahealth/CrisperWhisper)**
- **[丹麦语 / Danish](https://huggingface.co/sam8000/whisper-large-v3-turbo-danish-denmark)**
- **[印度英语 / English-Indian](https://huggingface.co/Tejveer12/Indian-Accent-English-Whisper-Finetuned)**
- **[英语 v3.5 / Engish-v3.5](https://huggingface.co/distil-whisper/distil-large-v3.5)**
- **[法语 / French](https://huggingface.co/bofenghuang/whisper-large-v3-french-distil-dec16)**
- **[瑞士德语 / German-Swiss](https://huggingface.co/Flurin17/whisper-large-v3-turbo-swiss-german)**
- **[德语 / German](https://huggingface.co/primeline/whisper-large-v3-turbo-german)**
- **[希腊语 / Greek](https://huggingface.co/sam8000/whisper-large-v3-turbo-greek-greece)**
- **[意大利语 / Italian](https://huggingface.co/bofenghuang/whisper-large-v3-distil-it-v0.2)**
- **[日语-动漫 / Japanese-Anime](https://huggingface.co/efwkjn/whisper-ja-anime-v0.3)**
- **[日语 / Japanese](https://huggingface.co/hhim8826/whisper-large-v3-turbo-ja)**
- **[韩语 / Korean](https://huggingface.co/ghost613/whisper-large-v3-turbo-korean)**
- **[马来语 / Malaysian](https://huggingface.co/mesolitica/Malaysian-whisper-large-v3-turbo-v3)**
- **[波斯语 / Persian](https://huggingface.co/MohammadGholizadeh/whisper-large-v3-persian-common-voice-17)**
- **[波兰语 / Polish](https://huggingface.co/Aspik101/distil-whisper-large-v3-pl)**
- **[葡萄牙语 / Portuguese](https://huggingface.co/freds0/distil-whisper-large-v3-ptbr)**
- **[俄语 / Russian](https://huggingface.co/dvislobokov/whisper-large-v3-turbo-russian)**
- **[塞尔维亚语 / Serbian](https://huggingface.co/Sagicc/whisper-large-v3-sr-combined)**
- **[西班牙语 / Spanish](https://huggingface.co/Berly00/whisper-large-v3-spanish)**
- **[泰语 / Thai](https://huggingface.co/nectec/Pathumma-whisper-th-large-v3)**
- **[土耳其语 / Turkish](https://huggingface.co/selimc/whisper-large-v3-turbo-turkish)**
- **[乌尔都语 / Urdu](https://huggingface.co/urdu-asr/whisper-large-v3-ur)**
- **[越南语 / Vietnamese](https://huggingface.co/suzii/vi-whisper-large-v3-turbo-v1)**
### 🤖 翻译模型 (LLM) / Translation Models (LLM)
- **[Qwen-3-4B-Instruct-2507-Abliterated](https://huggingface.co/huihui-ai/Huihui-Qwen3-4B-Instruct-2507-abliterated)**
- **[Qwen-3-8B-Abliterated](https://huggingface.co/huihui-ai/Huihui-Qwen3-8B-abliterated-v2)**
- **[Hunyuan-MT-7B-Abliterated](https://huggingface.co/huihui-ai/Huihui-Hunyuan-MT-7B-abliterated)**
- **[Seed-X-PRO-7B](https://www.modelscope.cn/models/ByteDance-Seed/Seed-X-PPO-7B)**
---
## 🖥️ 硬件支持 / Hardware Support
<table>
<tr>
<td align="center"><strong>💻 中央處理器 (CPU)</strong></td>
<td align="center"><strong>🎮 圖形處理器 (GPU)</strong></td>
<td align="center"><strong>🧠 神經網路處理單元 (NPU)</strong></td>
</tr>
<tr>
<td valign="top">
<ul>
<li>Apple Silicon</li>
<li>AMD</li>
<li>ARM</li>
<li>Intel</li>
</ul>
</td>
<td valign="top">
<ul>
<li>Apple CoreML</li>
<li>AMD ROCm</li>
<li>Intel OpenVINO</li>
<li>NVIDIA CUDA</li>
<li>Windows DirectML</li>
</ul>
</td>
<td valign="top">
<ul>
<li>Apple CoreML</li>
<li>AMD Ryzen-VitisAI</li>
<li>Intel OpenVINO</li>
</ul>
</td>
</tr>
</table>
---
## 📊 性能基准测试 / Performance Benchmarks
*测试条件 / Test Conditions: Ubuntu 24.04, Intel i3-12300, 7602 秒视频*
| 操作系统 (OS) | 后端 (Backend) | 降噪器 (Denoiser) | VAD | 语音识别 (ASR) | 大语言模型 (LLM) | 实时率<br>(Real-Time Factor) |
|:---:|:---:|:---:|:---:|:---:|:---:|:---:|
| Ubuntu-24.04 | CPU i3-12300 | - | Silero | SenseVoiceSmall | - | **0.08** |
| Ubuntu-24.04 | CPU i3-12300 | GTCRN | Silero | SenseVoiceSmall | Qwen2.5-7B-Instruct | **0.50** |
| Ubuntu-24.04 | CPU i3-12300 | GTCRN | FSMN | SenseVoiceSmall | - | **0.054** |
| Ubuntu-24.04 | CPU i3-12300 | ZipEnhancer | FSMN | SenseVoiceSmall | - | **0.39** |
| Ubuntu-24.04 | CPU i3-12300 | GTCRN | Silero | Whisper-Large-V3 | - | **0.20** |
| Ubuntu-24.04 | CPU i3-12300 | GTCRN | FSMN | Whisper-Large-V3-Turbo | - | **0.148** |
---
## 🛠️ 问题排查 / Troubleshooting
### 常见问题 / Common Issues
- **Silero VAD 错误 / Silero VAD Error**: 首次运行时只需重启应用程序 / Simply restart the application on first run
- **libc++ 错误 (Linux) / libc++ Error (Linux)**:
```bash
sudo apt update
sudo apt install libc++1
```
- **苹果芯片 / Apple Silicon**: 请避免安装 `onnxruntime-openvino`,因为它会导致错误 / Avoid installing `onnxruntime-openvino` as it will cause errors
---
## 📋 更新历史 / Update History
### 🆕 **2025/9/19** - 重大更新 / Major Release
- ✅ **新增 ASR / Added ASR**:
- 28 个地区微调的 Whisper 模型
- 28 region fine-tuned Whisper models
- ✅ **新增降噪器 / Added Denoiser**: MossFormer2_SE_48K
- ✅ **新增 LLM 模型 / Added LLM Models**:
- Qwen3-4B-Instruct-2507-abliterated
- Qwen3-8B-abliterated-v2
- Hunyuan-MT-7B-abliterated
- Seed-X-PRO-7B
- ✅ **性能改进 / Performance Improvements**:
- 为类 Whisper 的 ASR 模型应用了束搜索(Beam Search)和重复惩罚(Repeat Penalty)
- 应用 ONNX Runtime IOBinding 实现最大加速(比常规 ort_session.run() 快 10%以上)
- 支持单次推理处理 20 秒的音频片段
- 改进了多线程性能
- Applied Beam Search & Repeat Penalty for Whisper-like ASR models
- Applied ONNX Runtime IOBinding for maximum speed up (10%+ faster than normal ort_session.run())
- Support for 20 seconds audio segment per single run inference
- Improved multi-threads performance
- ✅ **硬件支持扩展 / Hardware Support Expansion**:
- AMD-ROCm 执行提供程序 / Execution Provider
- AMD-MIGraphX 执行提供程序 / Execution Provider
- NVIDIA TensorRTX 执行提供程序 / Execution Provider
- *(必须先配置环境,否则无法工作 / Must config the env first or it will not work)*
- ✅ **准确性改进 / Accuracy Improvements**:
- SenseVoice
- Paraformer
- FireRedASR
- Dolphin
- ZipEnhancer
- MossFormerGAN_SE_16K
- NVIDIA-NeMo-VAD
- ✅ **速度改进 / Speed Improvements**:
- MelBandRoformer (通过转换为单声道提升速度 / speed boost by converting to mono channel)
- ❌ **移除的模型 / Removed Models**:
- FSMN-VAD
- Qwen3-4B-Official
- Qwen3-8B-Official
- Gemma3-4B-it
- Gemma3-12B-it
- InternLM3
- Phi-4-Instruct
### **2025/7/5** - 降噪增强 / Noise Reduction Enhancement
- ✅ **新增降噪模型 / Added noise reduction model**: MossFormerGAN_SE_16K
### **2025/6/11** - VAD 模型扩展 / VAD Models Expansion
- ✅ **新增 VAD 模型 / Added VAD Models**:
- HumAware-VAD
- NVIDIA-NeMo-VAD
- TEN-VAD
### **2025/6/3** - 亚洲语言支持 / Asian Language Support
- ✅ **新增 Dolphin ASR 模型以支持亚洲语言 / Added Dolphin ASR model** to support Asian languages
### **2025/5/13** - GPU 加速 / GPU Acceleration
- ✅ **新增 Float16/32 ASR 模型以支持 CUDA/DirectML GPU / Added Float16/32 ASR models** to support CUDA/DirectML GPU usage
- ✅ **GPU 性能 / GPU Performance**: 这些模型可以实现超过 99% 的 GPU 算子部署 / These models can achieve >99% GPU operator deployment
### **2025/5/9** - 主要功能发布 / Major Feature Release
- ✅ **灵活性改进 / Flexibility Improvements**:
- 新增不使用 VAD(语音活动检测)的选项 / Added option to **not use** VAD (Voice Activity Detection)
- ✅ **新增模型 / Added Models**:
- **降噪 / Noise reduction**: MelBandRoformer
- **ASR**: CrisperWhisper
- **ASR**: Whisper-Large-v3.5-Distil (英语微调 / English fine-tuned)
- **ASR**: FireRedASR-AED-L (支持中文及方言 / Chinese + dialects support)
- **三个日语动漫微调的 Whisper 模型 / Three Japanese anime fine-tuned Whisper models**
- ✅ **性能优化 / Performance Optimizations**:
- 移除 IPEX-LLM 框架以提升整体性能 / Removed IPEX-LLM framework to enhance overall performance
- 取消 LLM 量化选项,统一使用 **Q4F32** 格式 / Cancelled LLM quantization options, standardized on **Q4F32** format
- Whisper 系列推理速度提升 10% 以上 / Improved **Whisper** series inference speed by over 10%
- ✅ **准确性改进 / Accuracy Improvements**:
- 提升 **FSMN-VAD** 准确率 / Improved **FSMN-VAD** accuracy
- 提升 **Paraformer** 识别准确率 / Improved **Paraformer** recognition accuracy
- 提升 **SenseVoice** 识别准确率 / Improved **SenseVoice** recognition accuracy
- ✅ **LLM 支持 ONNX Runtime 100% GPU 算子部署 / LLM Support with ONNX Runtime 100% GPU operator deployment**:
- Qwen3-4B/8B
- InternLM3-8B
- Phi-4-mini-Instruct
- Gemma3-4B/12B-it
- ✅ **硬件支持扩展 / Hardware Support Expansion**:
- **Intel OpenVINO**
- **NVIDIA CUDA GPU**
- **Windows DirectML GPU** (支持集成显卡和独立显卡 / supports integrated and discrete GPUs)
---
## 🗺️ 路线图 / Roadmap
- [ ] **视频超分 / [Video Upscaling](https://github.com/ByteDance-Seed/SeedVR/tree/main)** - 提升分辨率 / Enhance resolution
- [ ] **实时播放器 / Real-time Player** - 实时转录和翻译 / Live transcription and translation
|
tinsamine535/blockassist
|
tinsamine535
| 2025-09-25T05:37:03Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"scaly stinging mongoose",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-18T17:11:55Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- scaly stinging mongoose
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
bahree/london-historical-slm
|
bahree
| 2025-09-25T05:36:31Z | 87 | 0 |
transformers
|
[
"transformers",
"safetensors",
"gpt2",
"text-generation",
"historical",
"london",
"slm",
"small-language-model",
"history",
"english",
"en",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
| 2025-09-14T18:49:33Z |
---
license: mit
library_name: transformers
pipeline_tag: text-generation
language:
- en
tags:
- gpt2
- historical
- london
- slm
- small-language-model
- text-generation
- history
- english
- safetensors
---
# London Historical LLM – Small Language Model (SLM)
A compact GPT-2 Small model (~117M params) **trained from scratch** on historical London texts (1500–1850). Fast to run on CPU, and supports NVIDIA (CUDA) and AMD (ROCm) GPUs.
**Note**:
* This model was **trained from scratch** - not fine-tuned from existing models.
* This page includes simple **virtual-env setup**, **install choices for CPU/CUDA/ROCm**, and an **auto-device inference** example so anyone can get going quickly.
---
## 🔎 Model Description
This is a **Small Language Model (SLM)** version of the London Historical LLM, **trained from scratch** using GPT-2 Small architecture on historical London texts with a custom historical tokenizer. The model was built from the ground up, not fine-tuned from existing models.
### Key Features
- ~117M parameters (vs ~354M in the full model)
- Custom historical tokenizer (≈30k vocab)
- London-specific context awareness and historical language patterns (e.g., *thou, thee, hath*)
- Lower memory footprint and faster inference on commodity hardware
- **Trained from scratch** - not fine-tuned from existing models
---
## Repository
The complete source code, training scripts, and documentation for this model are available on GitHub:
** [https://github.com/bahree/helloLondon](https://github.com/bahree/helloLondon)**
This repository includes:
- Complete data collection pipeline for 1500-1850 historical English
- Custom tokenizer optimized for historical text
- Training infrastructure with GPU optimization
- Evaluation and deployment tools
- Comprehensive documentation and examples
### Quick Start with Repository
```bash
git clone https://github.com/bahree/helloLondon.git
cd helloLondon
python 06_inference/test_published_models.py --model_type slm
```
---
## 🧪 Intended Use & Limitations
**Use cases:** historical-style narrative generation, prompt-based exploration of London themes (1500–1850), creative writing aids.
**Limitations:** may produce anachronisms or historically inaccurate statements; smaller models have less complex reasoning than larger LLMs. Validate outputs before downstream use.
---
## 🐍 Set up a virtual environment (Linux/macOS/Windows)
> Virtual environments isolate project dependencies. Official Python docs: `venv`.
**Check Python & pip**
```bash
# Linux/macOS
python3 --version && python3 -m pip --version
```
```powershell
# Windows (PowerShell)
python --version; python -m pip --version
```
**Create the env**
```bash
# Linux/macOS
python3 -m venv helloLondon
```
```powershell
# Windows (PowerShell)
python -m venv helloLondon
```
```cmd
:: Windows (Command Prompt)
python -m venv helloLondon
```
> **Note**: You can name your virtual environment anything you like, e.g., `.venv`, `my_env`, `london_env`.
**Activate**
```bash
# Linux/macOS
source helloLondon/bin/activate
```
```powershell
# Windows (PowerShell)
.\helloLondon\Scripts\Activate.ps1
```
```cmd
:: Windows (CMD)
.\helloLondon\Scripts\activate.bat
```
> If PowerShell blocks activation (*"running scripts is disabled"*), set the policy then retry activation:
```powershell
Set-ExecutionPolicy -Scope CurrentUser -ExecutionPolicy RemoteSigned
# or just for this session:
Set-ExecutionPolicy -Scope Process -ExecutionPolicy Bypass
```
---
## 📦 Install libraries
Upgrade basics, then install Hugging Face libs:
```bash
python -m pip install -U pip setuptools wheel
python -m pip install "transformers[torch]" accelerate safetensors
```
---
## 🚀 Inference (auto-detect device)
This snippet picks the best device (CUDA/ROCm if available, else CPU) and uses sensible generation defaults for this SLM.
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_id = "bahree/london-historical-slm"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
device = "cuda" if torch.cuda.is_available() else "cpu"
model = model.to(device)
prompt = "In the year 1834, I walked through the streets of London and witnessed"
inputs = tokenizer(prompt, return_tensors="pt").to(device)
outputs = model.generate(
inputs["input_ids"],
max_new_tokens=50,
do_sample=True,
temperature=0.8,
top_p=0.95,
top_k=40,
repetition_penalty=1.2,
no_repeat_ngram_size=3,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=tokenizer.eos_token_id,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
---
## 📖 **Sample Output**
**Prompt:** "In the year 1834, I walked through the streets of London and witnessed"
**Generated Text:**
> "In the year 1834, I walked through the streets of London and witnessed a scene in which some of those who had no inclination to come in contact with him took part in his discourse. It was on this occasion that I perceived that he had been engaged in some new business connected with the house, but for some days it had not taken place, nor did he appear so desirous of pursuing any further display of interest. The result was, however, that if he came in contact witli any one else in company with him he must be regarded as an old acquaintance or companion, and when he came to the point of leaving, I had no leisure to take up his abode. The same evening, having ram ##bled about the streets, I observed that the young man who had just arrived from a neighbouring village at the time, was enjoying himself at a certain hour, and I thought that he would sleep quietly until morning, when he said in a low voice — " You are coming. Miss — I have come from the West Indies. " Then my father bade me go into the shop, and bid me put on his spectacles, which he had in his hand; but he replied no: the room was empty, and he did not want to see what had passed. When I asked him the cause of all this conversation, he answered in the affirmative, and turned away, saying that as soon as the lad could recover, the sight of him might be renewed. " Well, Mr. , " said I, " you have got a little more of your wages, do you? " " No, sir, thank 'ee kindly, " returned the boy, " but we don 't want to pay the poor rates. We"
---
## 🧪 **Testing Your Model**
### **Quick Testing (Recommended First)**
```bash
# Test the published model with 10 automated prompts
python 06_inference/test_published_models.py --model_type slm
```
**What this does:**
- Loads model from `bahree/london-historical-slm`
- Tests 10 historical prompts automatically
- Shows model info (vocab size, parameters, etc.)
- Uses SLM-optimized generation parameters
- **No user interaction** - just runs and reports results
**Expected Output:**
```
🧪 Testing SLM Model: bahree/london-historical-slm
============================================================
📂 Loading model...
✅ Model loaded in 8.91 seconds
📊 Model Info:
Type: SLM
Description: Small Language Model (117M parameters)
Device: cuda
Vocabulary size: 30,000
Max length: 512
🎯 Testing generation with 10 prompts...
--- Test 1/10 ---
Prompt: In the year 1834, I walked through the streets of London and witnessed
Generated: . a scene in which some of those who had no inclination to come in contact with him took part in his discourse . It was on this occasion that I perceived that he had been engaged in some new business connected with the house , but for some days it had not taken place , nor did he appear so desirous of pursuing any further display of interest . The result was , however , that if he came in contact witli any one else in company with him he must be regarded as an old acquaintance or companion , and when he came to the point of leaving , I had no leisure to take up his abode . The same evening , having ram ##bled about the streets , I observed that the young man who had just arrived from a neighbouring village at the time , was enjoying himself at a certain hour , and I thought that he would sleep quietly until morning , when he said in a low voice — " You are coming . Miss — I have come from the West Indies . " Then my father bade me go into the shop , and bid me put on his spectacles , which he had in his hand ; but he replied no : the room was empty , and he did not want to see what had passed . When I asked him the cause of all this conversation , he answered in the affirmative , and turned away , saying that as soon as the lad could recover , the sight of him might be renewed . " Well , Mr . , " said I , " you have got a little more of your wages , do you ? " " No , sir , thank ' ee kindly , " returned the boy , " but we don ' t want to pay the poor rates . We
```
### **Interactive Testing (For Exploration)**
```bash
# Interactive mode for custom prompts
python 06_inference/inference_unified.py --published --model_type slm --interactive
# Single prompt test
python 06_inference/inference_unified.py --published --model_type slm --prompt "In the year 1834, I walked through the streets of London and witnessed"
```
---
## 🛠️ Training Details
* **Architecture:** Custom GPT-2 Small (built from scratch)
* **Parameters:** ~117M
* **Tokenizer:** Custom historical tokenizer (~30k vocab) with London-specific and historical tokens
* **Data:** Historical London corpus (1500-1850) with proper segmentation
* **Steps:** 4,000 steps (early stopping for SLM)
* **Final Training Loss:** ~3.08 (good convergence)
* **Final Validation Loss:** ~3.67 (good generalization)
* **Training Time:** ~1.5 hours
* **Hardware:** 1× GPU training
* **Training Method:** **Train from scratch** using `04_training/train_model_slm.py`
---
## 🔤 Historical Tokenizer
* Compact 30k vocab targeting 1500–1850 English
* Tokens for **year/date/name/place/title**, plus **thames**, **westminster**, etc.; includes **thou/thee/hath/doth** style markers
---
## ⚠️ Troubleshooting
* **`ImportError: AutoModelForCausalLM requires the PyTorch library`**
→ Install PyTorch with the correct accelerator variant (see CPU/CUDA/ROCm above or use the official selector).
* **AMD GPU not used**
→ Ensure you installed a ROCm build and you're on Linux (`pip install ... --index-url https://download.pytorch.org/whl/rocmX.Y`). Verify with `torch.cuda.is_available()` and check the device name. ROCm wheels are Linux-only.
* **Running out of VRAM**
→ Try smaller batch/sequence lengths, or load with `device_map="auto"` via 🤗 Accelerate to offload layers to CPU/disk.
---
## 📚 Citation
If you use this model, please cite:
```bibtex
@misc{london-historical-slm,
title = {London Historical LLM - Small Language Model: A Compact GPT-2 for Historical Text Generation},
author = {Amit Bahree},
year = {2025},
url = {https://huggingface.co/bahree/london-historical-slm}
}
```
---
## 🧾 License
MIT (see [LICENSE](https://huggingface.co/bahree/london-historical-slm/blob/main/LICENSE) in repo).
|
hdvc3395/blockassist
|
hdvc3395
| 2025-09-25T05:35:36Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"leggy muscular alpaca",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-20T12:19:36Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- leggy muscular alpaca
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
tv611527/blockassist
|
tv611527
| 2025-09-25T05:35:10Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"fast peaceful mouse",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:51:36Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- fast peaceful mouse
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
manskeolin73/blockassist
|
manskeolin73
| 2025-09-25T05:34:38Z | 0 | 0 | null |
[
"gensyn",
"blockassist",
"gensyn-blockassist",
"minecraft",
"chattering quiet bee",
"arxiv:2504.07091",
"region:us"
] | null | 2025-09-21T03:51:04Z |
---
tags:
- gensyn
- blockassist
- gensyn-blockassist
- minecraft
- chattering quiet bee
---
# Gensyn BlockAssist
Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
|
Subsets and Splits
Filtered Qwen2.5 Distill Models
Identifies specific configurations of models by filtering cards that contain 'distill', 'qwen2.5', '7b' while excluding certain base models and incorrect model ID patterns, uncovering unique model variants.
Filtered Model Cards Count
Finds the count of entries with specific card details that include 'distill', 'qwen2.5', '7b' but exclude certain base models, revealing valuable insights about the dataset's content distribution.
Filtered Distill Qwen 7B Models
Filters for specific card entries containing 'distill', 'qwen', and '7b', excluding certain strings and patterns, to identify relevant model configurations.
Filtered Qwen-7b Model Cards
The query performs a detailed filtering based on specific keywords and excludes certain entries, which could be useful for identifying a specific subset of cards but does not provide deeper insights or trends.
Filtered Qwen 7B Model Cards
The query filters for specific terms related to "distilled" or "distill", "qwen", and "7b" in the 'card' column but excludes certain base models, providing a limited set of entries for further inspection.
Qwen 7B Distilled Models
The query provides a basic filtering of records to find specific card names that include keywords related to distilled Qwen 7b models, excluding a particular base model, which gives limited insight but helps in focusing on relevant entries.
Qwen 7B Distilled Model Cards
The query filters data based on specific keywords in the modelId and card fields, providing limited insight primarily useful for locating specific entries rather than revealing broad patterns or trends.
Qwen 7B Distilled Models
Finds all entries containing the terms 'distilled', 'qwen', and '7b' in a case-insensitive manner, providing a filtered set of records but without deeper analysis.
Distilled Qwen 7B Models
The query filters for specific model IDs containing 'distilled', 'qwen', and '7b', providing a basic retrieval of relevant entries but without deeper analysis or insight.
Filtered Model Cards with Distill Qwen2.
Filters and retrieves records containing specific keywords in the card description while excluding certain phrases, providing a basic count of relevant entries.
Filtered Model Cards with Distill Qwen 7
The query filters specific variations of card descriptions containing 'distill', 'qwen', and '7b' while excluding a particular base model, providing limited but specific data retrieval.
Distill Qwen 7B Model Cards
The query filters and retrieves rows where the 'card' column contains specific keywords ('distill', 'qwen', and '7b'), providing a basic filter result that can help in identifying specific entries.