TinyLlama-1.1bee π
As we feverishly hit the refresh button on hf.co's homepage, on the hunt for the newest waifu chatbot to grace the AI stage, an epiphany struck us like a bee sting. What could we offer to the hive-mind of the community? The answer was as clear as honeyβbeekeeping, naturally. And thus, this un-bee-lievable model was born.
Details
This model is a fine-tuned version of PY007/TinyLlama-1.1B-intermediate-step-240k-503b on the BEE-spoke-data/bees-internal
dataset.
It achieves the following results on the evaluation set:
- Loss: 2.4285
- Accuracy: 0.4969
***** eval metrics *****
eval_accuracy = 0.4972
eval_loss = 2.4283
eval_runtime = 0:00:53.12
eval_samples = 239
eval_samples_per_second = 4.499
eval_steps_per_second = 1.129
perplexity = 11.3391
π Intended Uses & Limitations π
Intended Uses:
- Educational Engagement: Whether you're a novice beekeeper, an enthusiast, or someone just looking to understand the buzz around bees, this model aims to serve as an informative and entertaining resource.
- General Queries: Have questions about hive management, bee species, or honey extraction? Feel free to consult the model for general insights.
- Academic & Research Inspiration: If you're diving into the world of apiculture studies or environmental science, our model could offer some preliminary insights and ideas.
Limitations:
- Not a Beekeeping Expert: As much as we admire bees and their hard work, this model is not a certified apiculturist. Please consult professional beekeeping resources or experts for serious decisions related to hive management, bee health, and honey production.
- Licensing: Apache-2.0, following TinyLlama
- Infallibility: Our model can err, just like any other piece of technology (or bee). Always double-check the information before applying it to your own hive or research.
- Ethical Constraints: This model may not be used for any illegal or unethical activities, including but not limited to: bioterrorism & standard terrorism, harassment, or spreading disinformation.
Training and evaluation data
While the full dataset is not yet complete and therefore not yet released for "safety reasons", you can check out a preliminary sample at: bees-v0
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 4
- eval_batch_size: 4
- seed: 80085
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.03
- num_epochs: 2.0
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 29.15 |
ARC (25-shot) | 30.55 |
HellaSwag (10-shot) | 51.8 |
MMLU (5-shot) | 24.25 |
TruthfulQA (0-shot) | 39.01 |
Winogrande (5-shot) | 54.46 |
GSM8K (5-shot) | 0.23 |
DROP (3-shot) | 3.74 |
- Downloads last month
- 18
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
Model tree for BEE-spoke-data/TinyLlama-1.1bee
Dataset used to train BEE-spoke-data/TinyLlama-1.1bee
Collection including BEE-spoke-data/TinyLlama-1.1bee
Collection
models fine-tuned to be knowledgeable about apiary practice
β’
6 items
β’
Updated
β’
1