modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-07-16 00:42:46
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
522 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-07-16 00:42:16
card
stringlengths
11
1.01M
prince-canuma/Damysus-2.7B-Chat-GGUF
prince-canuma
2024-02-17T10:30:24Z
26
0
transformers
[ "transformers", "gguf", "phi", "text-generation", "nlp", "phi-2", "instruct", "conversational", "custom_code", "en", "dataset:Open-Orca/SlimOrca", "dataset:prince-canuma/TinyOrca", "base_model:microsoft/phi-2", "base_model:quantized:microsoft/phi-2", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-02-15T13:19:37Z
--- language: - en license: mit library_name: transformers tags: - nlp - phi - phi-2 - instruct base_model: - microsoft/phi-2 datasets: - Open-Orca/SlimOrca - prince-canuma/TinyOrca model-index: - name: Damysus-2.7B-Chat results: - task: type: text-generation metrics: - name: Average type: Average value: 60.49 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: ARC (25-shot) type: ai2_arc metrics: - name: Accuracy Norm type: acc_norm value: 59.81 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: Hellaswag (10-shot) type: Hellaswag metrics: - name: Accuracy Norm type: acc value: 74.52 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: MMLU (5-shot) type: MMLU metrics: - name: Accuracy type: acc value: 56.33 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: Truthful QA type: Truthful_QA metrics: - name: Multi-true type: mc2 value: 46.74 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: Winogrande (5-shot) type: Winogrande metrics: - name: Accuracy type: acc value: 75.06 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: GSM8K (5-shot) type: GSM8K metrics: - name: Accuracy type: acc value: 50.64 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard --- # Model Summary <img src="Damysus.png" width="500" alt="Damysus - the fastest giant"/> <!-- Provide a quick summary of what the model is/does. --> This model is a GGUF version of [Damysus-2.7B-Chat](https://huggingface.co/prince-canuma/Damysus-2.7B-Chat). ## Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [Prince Canuma](https://huggingface.co/prince-canuma) - **Model type:** Transformer - **License:** MIT - **Finetuned from model:** microsoft/phi-2 ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> You can use this model to build local/cloud RAG applications. It can serve as the: - Answer synthesizer, - Summarizer, - Or query rewriter model. ### Limitations This model inherits some of the base model's limitations, such as: - Generate Inaccurate Code and Facts: The model may produce incorrect code snippets and statements. Users should treat these outputs as suggestions or starting points, not as definitive or accurate solutions. - Limited Scope for code: Majority of Phi-2 training data is based in Python and use common packages such as "typing, math, random, collections, datetime, itertools". If the model generates Python scripts that utilize other packages or scripts in other languages, we strongly recommend users manually verify all API uses. - Language Limitations: The model is primarily designed to understand standard English. Informal English, slang, or any other languages might pose challenges to its comprehension, leading to potential misinterpretations or errors in response. ### On the command line, including multiple files at once I recommend using the `huggingface-hub` Python library: ```shell pip3 install huggingface-hub ``` Then you can download any individual model file to the current directory, at high speed, with a command like this: ```shell huggingface-cli download prince-canuma/Damysus-2.7B-Chat-GGUF Damysus-2.7B-Chat.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False ``` <details> <summary>More advanced huggingface-cli download usage (click to read)</summary> You can also download multiple files at once with a pattern: ```shell huggingface-cli download prince-canuma/Damysus-2.7B-Chat-GGUF --local-dir . --local-dir-use-symlinks False --include='*Q4_K*gguf' ``` For more documentation on downloading with `huggingface-cli`, please see: [HF -> Hub Python Library -> Download files -> Download from the CLI](https://huggingface.co/docs/huggingface_hub/guides/download#download-from-the-cli). To accelerate downloads on fast connections (1Gbit/s or higher), install `hf_transfer`: ```shell pip3 install hf_transfer ``` And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`: ```shell HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download prince-canuma/Damysus-2.7B-Chat-GGUF Damysus-2.7B-Chat.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False ``` Windows Command Line users: You can set the environment variable by running `set HF_HUB_ENABLE_HF_TRANSFER=1` before the download command. </details> <!-- README_GGUF.md-how-to-download end --> <!-- README_GGUF.md-how-to-run start --> ## Example `llama.cpp` command Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later. ```shell !./main -m ../Damysus-2.7B-Chat-GGUF/Damysus-2.7B-Chat.Q4_K_M.gguf \ --color -c 2048 --temp 0 \ --prompt "<|im_start|>system\nYou are a helpful assistant. Please keep your answers short.<|im_end|>\n<|im_start|>user\nCount to ten<|im_end|>\n" \ -n 256 --in-suffix "<|im_start|>assistant\n" -r "User:" -e --verbose-prompt ``` or ```shell !./main -m ../Damysus-2.7B-Chat-GGUF/Damysus-2.7B-Chat.Q4_K_M.gguf \ --color -c 2048 --temp 0 \ -p "You are a helpful assistant. Please keep your answers short." -n 256 --in-suffix "<|im_start|>assistant\n" \ -r "User:" -e --verbose-prompt -cml ``` - `-ngl N` offload N number of layers to GPU. Remove it if you don't have GPU acceleration. - `-c 2048` set desired sequence length. For extended sequence models - eg 8K, 16K, 32K - the necessary RoPE scaling parameters are read from the GGUF file and set by llama.cpp automatically. Note that longer sequence lengths require much more resources, so you may need to reduce this value. - Add `-i -ins` or `-cml` argument for interactive chat-style conversation. For other parameters and how to use them, please refer to [the llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/main/README.md) or run: ```shell !./main --help ``` ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> I used [SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca) dataset, a new curated subset of our OpenOrca data. In the course of this study, the [SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca) dataset was used, representing a meticulously curated subset derived from the broader OpenOrca dataset. This release provides an efficient means of reaching performance on-par with using larger slices of the [OpenOrca](https://huggingface.co/datasets/Open-Orca/OpenOrca), while only including ~500k GPT-4 completions. Subsequently, two distinct subsets were crafted, comprising 102,000 and 1,000 samples, denoted as: - [prince-canuma/SmallOrca](https://huggingface.co/datasets/prince-canuma/SmallOrca) - [prince-canuma/TinyOrca](https://huggingface.co/datasets/prince-canuma/TinyOrca) Although experimentation was conducted with both datasets, optimal results were achieved through fine-tuning on a modest set of 200 samples. Notably, the investigation revealed that augmenting the training data beyond this threshold predominantly enhanced the model's proficiency in generating Chain-of-Thought responses. However, it is imperative to note that the preference for Chain-of-Thought responses may not be universally applicable. Particularly in scenarios like the RAG setup, succinct answers to prompts are often favored, especially for straightforward queries. ### Training Procedure #### Preprocessing 1. Convert dataset to chatML format 2. Remove all samples with more than 2048 tokens (Phi-2 context size) 3. Mask instructions (System and User) at training time. #### LoRA Config - **lora_alpha:** 128, - **lora_dropout:** 0.05, - **r:** 256, - **bias:** "none", - **target_modules:** "all-linear", - **task_type:** "CAUSAL_LM", #### Training Hyperparameters - **Training regime:** bf16 mixed precision, <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> - **max_steps:** 100, - **per_device_train_batch_size:** 2, - **gradient_accumulation_steps:** 2, - **optim:** "adamw_torch_fused", - **learning_rate:** 2e-4, - **max_grad_norm:** 0.3, - **warmup_ratio:** 0.03, - **lr_scheduler_type:** "constant", #### Trainer - **max_seq_length:** 1744, - **data_collator:** DataCollatorForCompletionOnlyLM ## Evaluation <img src="truthfulQA.png" width="800" alt="Damysus-2.7B-chat truthfulQA benchmark results"/> <!-- This section describes the evaluation protocols and provides the results. --> We evaluate models on 7 key benchmarks using the Eleuther AI Language Model Evaluation Harness , a unified framework to test generative language models on a large number of different evaluation tasks. - AI2 Reasoning Challenge (25-shot) - a set of grade-school science questions. - HellaSwag (10-shot) - a test of commonsense inference, which is easy for humans (~95%) but challenging for SOTA models. - MMLU (5-shot) - a test to measure a text model's multitask accuracy. The test covers 57 tasks including elementary mathematics, US history, computer science, law, and more. - TruthfulQA (0-shot) - a test to measure a model's propensity to reproduce falsehoods commonly found online. Note: TruthfulQA is technically a 6-shot task in the Harness because each example is prepended with 6 Q/A pairs, even in the 0-shot setting. - Winogrande (5-shot) - an adversarial and difficult Winograd benchmark at scale, for commonsense reasoning. - GSM8k (5-shot) - diverse grade school math word problems to measure a model's ability to solve multi-step mathematical reasoning problems. For all these evaluations, a higher score is a better score. We chose these benchmarks as they test a variety of reasoning and general knowledge across a wide variety of fields in 0-shot and few-shot settings. Read more [here](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). ### Results | Model | AVG | ARC | Hellaswag | MMLU | Truthful QA | Winogrande | GSM8K | |-------|--------:|------:|----------:|-----:|----------:|----------:|----------:| | [NousResearch/Nous-Puffin-70B](NousResearch/Nous-Puffin-70B) | 64.91 | 67.41 | 87.37 | 69.77 | 46.77 | 83.9 | 34.27 | | [TheBloke/Llama-2-70B-fp16](https://huggingface.co/TheBloke/Llama-2-70B-fp16) | 64.52 | 67.32 | 87.33 | 69.83 | 44.92 | 83.74 | 33.97 | | [NousResearch/Yarn-Mistral-7B-64k](https://huggingface.co/NousResearch/Yarn-Mistral-7b-64k) | 59.63 | 59.9 | 82.51 | 62.96 | 41.86 | 77.27 | 33.28 | | [Qwen1.5-4B-Chat](https://huggingface.co/Qwen/Qwen1.5-4B-Chat) | 46.79 | 43.26 | 69.73 | 55.55 | 44.79 | 64.96 | 2.43 | | [Microsoft/phi-2](https://huggingface.co/microsoft/phi-2) | 61.33 | 61.09 | 75.11 | 58.11 | 44.47 | 74.35 | 54.81 | | [Damysus-2.7B-Chat](https://huggingface.co/prince-canuma/Damysus-2.7B-Chat) (Ours) | 60.49 | 59.81 | 74.52 | 56.33 | **46.74** | **75.06** | 50.64 | ## Technical Specifications ### Compute Infrastructure - Modal Labs #### Hardware - OS: Linux - GPU: A10G #### Libraries - TRL - Transformers - PEFT - Datasets - Accelerate - torch - Wandb - Bitsandbytes - Plotly ## Future work I plan to explore the following tuning setups: - Function calling - DPO ## Citation <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** ```bibtex @misc{Damysus-2.7B-Chat, title={Damysus-2.7B-Chat} , author={Prince Canuma}, year={2024}, } ``` ```bibtex @misc{SlimOrca, title = {SlimOrca: An Open Dataset of GPT-4 Augmented FLAN Reasoning Traces, with Verification}, author = {Wing Lian and Guan Wang and Bleys Goodson and Eugene Pentland and Austin Cook and Chanvichet Vong and "Teknium"}, year = {2023}, publisher = {HuggingFace}, url = {https://https://huggingface.co/Open-Orca/SlimOrca} } ``` ```bibtex @misc{open-llm-leaderboard, author = {Edward Beeching and Clémentine Fourrier and Nathan Habib and Sheon Han and Nathan Lambert and Nazneen Rajani and Omar Sanseviero and Lewis Tunstall and Thomas Wolf}, title = {Open LLM Leaderboard}, year = {2023}, publisher = {Hugging Face}, howpublished = "\url{https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard}" } ```
Bajiyo/malayalam_imasc
Bajiyo
2024-02-17T10:29:47Z
4
0
transformers
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "generated_from_trainer", "base_model:openai/whisper-small", "base_model:finetune:openai/whisper-small", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2024-02-16T11:27:48Z
--- license: apache-2.0 base_model: openai/whisper-small tags: - generated_from_trainer metrics: - wer model-index: - name: malayalam_imasc results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # malayalam_imasc This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0305 - Wer: 21.4941 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 2000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.1047 | 0.26 | 500 | 0.0947 | 47.9746 | | 0.0583 | 0.52 | 1000 | 0.0538 | 32.2169 | | 0.0457 | 0.77 | 1500 | 0.0374 | 24.9007 | | 0.029 | 1.03 | 2000 | 0.0305 | 21.4941 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.1+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
wanggy1/classify-cognitive-distortions-llama2-13b
wanggy1
2024-02-17T10:24:30Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-02-17T10:24:30Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Ricosama/outputM1
Ricosama
2024-02-17T10:21:10Z
0
0
peft
[ "peft", "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:mistralai/Mixtral-8x7B-Instruct-v0.1", "base_model:adapter:mistralai/Mixtral-8x7B-Instruct-v0.1", "license:apache-2.0", "region:us" ]
null
2024-02-17T10:20:03Z
--- license: apache-2.0 library_name: peft tags: - trl - sft - generated_from_trainer base_model: mistralai/Mixtral-8x7B-Instruct-v0.1 model-index: - name: outputM1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # outputM1 This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2.5e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
washeed/audio-transcribe
washeed
2024-02-17T10:18:59Z
64
3
transformers
[ "transformers", "pytorch", "jax", "safetensors", "whisper", "automatic-speech-recognition", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2024-02-03T11:12:46Z
# to run simply install chocolatey run this on your cmd: ``` @"%SystemRoot%\System32\WindowsPowerShell\v1.0\powershell.exe" -NoProfile -InputFormat None -ExecutionPolicy Bypass -Command "[System.Net.ServicePointManager]::SecurityProtocol = 3072; iex ((New-Object System.Net.WebClient).DownloadString('https://community.chocolatey.org/install.ps1'))" && SET "PATH=%PATH%;%ALLUSERSPROFILE%\chocolatey\bin" ``` # after that install ffmpeg in your device using choco install by running this on cmd after: ``` choco install ffmpeg ``` # install dependencies in python IDE using: ``` pip install --upgrade pip pip install --upgrade git+https://github.com/huggingface/transformers.git accelerate datasets[audio] ``` # then lastly to inference the model: ``` import torch from transformers import AutoModelForSpeechSeq2Seq, AutoProcessor, pipeline device = "cuda:0" if torch.cuda.is_available() else "cpu" torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32 model_id = "washeed/audio-transcribe" model = AutoModelForSpeechSeq2Seq.from_pretrained( model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True, use_safetensors=True ) model.to(device) processor = AutoProcessor.from_pretrained(model_id) pipe = pipeline( "automatic-speech-recognition", model=model, tokenizer=processor.tokenizer, feature_extractor=processor.feature_extractor, max_new_tokens=128, chunk_length_s=30, batch_size=16, return_timestamps=True, torch_dtype=torch_dtype, device=device, ) result = pipe("audio.mp3") print(result["text"]) ``` # if you want to transcribe instead of translating just replace the : ```result = pipe("audio.mp3")``` # with ``` result = pipe("inference.mp3", generate_kwargs={"task": "transcribe"})```
Makengo/HasskakuForRunpod
Makengo
2024-02-17T10:18:08Z
0
0
null
[ "license:openrail", "region:us" ]
null
2024-01-25T12:09:57Z
--- license: openrail --- This is just for download to runpod This model is not mine
prince-canuma/Damysus-2.7B-Chat
prince-canuma
2024-02-17T10:13:39Z
59
4
transformers
[ "transformers", "safetensors", "phi", "text-generation", "nlp", "phi-2", "instruct", "conversational", "custom_code", "en", "dataset:Open-Orca/SlimOrca", "dataset:prince-canuma/TinyOrca", "base_model:microsoft/phi-2", "base_model:finetune:microsoft/phi-2", "license:mit", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-11T14:43:52Z
--- language: - en license: mit library_name: transformers tags: - nlp - phi - phi-2 - instruct base_model: - microsoft/phi-2 datasets: - Open-Orca/SlimOrca - prince-canuma/TinyOrca model-index: - name: Damysus-2.7B-Chat results: - task: type: text-generation metrics: - name: Average type: Average value: 60.49 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: ARC (25-shot) type: ai2_arc metrics: - name: Accuracy Norm type: acc_norm value: 59.81 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: Hellaswag (10-shot) type: Hellaswag metrics: - name: Accuracy Norm type: acc value: 74.52 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: MMLU (5-shot) type: MMLU metrics: - name: Accuracy type: acc value: 56.33 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: Truthful QA type: Truthful_QA metrics: - name: Multi-true type: mc2 value: 46.74 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: Winogrande (5-shot) type: Winogrande metrics: - name: Accuracy type: acc value: 75.06 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - task: type: text-generation dataset: name: GSM8K (5-shot) type: GSM8K metrics: - name: Accuracy type: acc value: 50.64 verified: true source: name: Open LLM Leaderboard url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard --- # Model Summary <img src="Damysus.png" width="500" alt="Damysus - the fastest giant"/> <!-- Provide a quick summary of what the model is/does. --> This model is a instruction-tuned version of Phi-2, a Transformer model with 2.7 billion parameters from Microsoft. The model has undergone further training to better follow specific user instructions, enhancing its ability to perform tasks as directed and improve its interaction with users. This additional training helps the model to understand context better, generate more accurate and relevant responses, and adapt to a wide range of language-based tasks such as: - Questions and Answers, - Data Extraction, - Structured Outputs (i.e., JSON outputs), - And providing explanations, ## Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [Prince Canuma](https://huggingface.co/prince-canuma) - **Model type:** Transformer - **License:** MIT - **Finetuned from model:** microsoft/phi-2 ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> You can use this model to build local/cloud RAG applications. It can serve as the: - Answer synthesizer, - Summarizer, - Or query rewriter model. ### Limitations This model inherits some of the base model's limitations, such as: - Generate Inaccurate Code and Facts: The model may produce incorrect code snippets and statements. Users should treat these outputs as suggestions or starting points, not as definitive or accurate solutions. - Limited Scope for code: Majority of Phi-2 training data is based in Python and use common packages such as "typing, math, random, collections, datetime, itertools". If the model generates Python scripts that utilize other packages or scripts in other languages, we strongly recommend users manually verify all API uses. - Language Limitations: The model is primarily designed to understand standard English. Informal English, slang, or any other languages might pose challenges to its comprehension, leading to potential misinterpretations or errors in response. ## How to Get Started with the Model Use the code below to get started with the model. ```python from transformers import pipeline, Conversation chatbot = pipeline("conversational", model="prince-canuma/Damysus-2.7B-Chat") conversation = Conversation("I'm looking for a movie - what's your favourite one?") output = chatbot(conversation) print(output) ``` Or you can instatiate the model and tokenizer directly ```python from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("prince-canuma/Damysus-2.7B-Chat") model = AutoModelForCausalLM.from_pretrained("prince-canuma/Damysus-2.7B-Chat") inputs = tokenizer.apply_chat_template( [ {"content":"You are an helpful AI assistant","role":"system"}, {"content":"I'm looking for a movie - what's your favourite one?","role":"user"}, ], add_generation_prompt=True, return_tensors="pt", ).to("cuda") outputs = model.generate(inputs, do_sample=False, max_new_tokens=256) input_length = inputs.shape[1] print(tokenizer.batch_decode(outputs[:, input_length:], skip_special_tokens=True)[0]) ``` Output: ```shell My favorite movie is "The Shawshank Redemption." It's a powerful and inspiring story about hope, friendship, and redemption. The performances by Tim Robbins and Morgan Freeman are exceptional, and the film's themes and messages are timeless. I highly recommend it to anyone who enjoys a well-crafted and emotionally engaging story. ``` ### Structured Output ```python from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("prince-canuma/Damysus-2.7B-Chat") model = AutoModelForCausalLM.from_pretrained("prince-canuma/Damysus-2.7B-Chat") inputs = tokenizer.apply_chat_template( [ {"content":"You are a Robot that ONLY outputs JSON. Use this structure: {'entities': [{'type':..., 'name':...}]}.","role":"system"}, {"content":""""Extract the entities of type 'technology' and 'file_type' in JSON format from the following passage: AI is a transformative force in document processing employing technologies such as 'Machine Learning (ML), Natural Language Processing (NLP) and Optical Character Recognition (OCR) to understand, interpret, and summarize text. These technologies enhance accuracy, increase efficiency, and allow you and your company to process high volumes of data in short amount of time. For instance, you can easily extract key points and summarize a large PDF document (i.e., 500 pages) in just a few seconds.""", "role":"user"}, ], add_generation_prompt=True, return_tensors="pt", ).to("cuda") outputs = model.generate(inputs, do_sample=False, max_new_tokens=256) input_length = inputs.shape[1] print(tokenizer.batch_decode(outputs[:, input_length:], skip_special_tokens=True)[0]) ``` Output: ```json { "entities": [ { "type": "technology", "name": "Machine Learning (ML)" }, { "type": "technology", "name": "Natural Language Processing (NLP)" }, { "type": "technology", "name": "Optical Character Recognition (OCR)" }, { "type": "file_type", "name": "PDF" }, ] } ``` ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> I used [SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca) dataset, a new curated subset of our OpenOrca data. In the course of this study, the [SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca) dataset was used, representing a meticulously curated subset derived from the broader OpenOrca dataset. This release provides an efficient means of reaching performance on-par with using larger slices of the [OpenOrca](https://huggingface.co/datasets/Open-Orca/OpenOrca), while only including ~500k GPT-4 completions. Subsequently, two distinct subsets were crafted, comprising 102,000 and 1,000 samples, denoted as: - [prince-canuma/SmallOrca](https://huggingface.co/datasets/prince-canuma/SmallOrca) - [prince-canuma/TinyOrca](https://huggingface.co/datasets/prince-canuma/TinyOrca) Although experimentation was conducted with both datasets, optimal results were achieved through fine-tuning on a modest set of 200 samples. Notably, the investigation revealed that augmenting the training data beyond this threshold predominantly enhanced the model's proficiency in generating Chain-of-Thought responses. However, it is imperative to note that the preference for Chain-of-Thought responses may not be universally applicable. Particularly in scenarios like the RAG setup, succinct answers to prompts are often favored, especially for straightforward queries. ### Training Procedure #### Preprocessing 1. Convert dataset to chatML format 2. Remove all samples with more than 2048 tokens (Phi-2 context size) 3. Mask instructions (System and User) at training time. #### LoRA Config - **lora_alpha:** 128, - **lora_dropout:** 0.05, - **r:** 256, - **bias:** "none", - **target_modules:** "all-linear", - **task_type:** "CAUSAL_LM", #### Training Hyperparameters - **Training regime:** bf16 mixed precision, <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> - **max_steps:** 100, - **per_device_train_batch_size:** 2, - **gradient_accumulation_steps:** 2, - **optim:** "adamw_torch_fused", - **learning_rate:** 2e-4, - **max_grad_norm:** 0.3, - **warmup_ratio:** 0.03, - **lr_scheduler_type:** "constant", #### Trainer - **max_seq_length:** 1744, - **data_collator:** DataCollatorForCompletionOnlyLM ## Evaluation <img src="truthfulQA.png" width="800" alt="Damysus-2.7B-chat truthfulQA benchmark results"/> <!-- This section describes the evaluation protocols and provides the results. --> We evaluate models on 7 key benchmarks using the Eleuther AI Language Model Evaluation Harness , a unified framework to test generative language models on a large number of different evaluation tasks. - AI2 Reasoning Challenge (25-shot) - a set of grade-school science questions. - HellaSwag (10-shot) - a test of commonsense inference, which is easy for humans (~95%) but challenging for SOTA models. - MMLU (5-shot) - a test to measure a text model's multitask accuracy. The test covers 57 tasks including elementary mathematics, US history, computer science, law, and more. - TruthfulQA (0-shot) - a test to measure a model's propensity to reproduce falsehoods commonly found online. Note: TruthfulQA is technically a 6-shot task in the Harness because each example is prepended with 6 Q/A pairs, even in the 0-shot setting. - Winogrande (5-shot) - an adversarial and difficult Winograd benchmark at scale, for commonsense reasoning. - GSM8k (5-shot) - diverse grade school math word problems to measure a model's ability to solve multi-step mathematical reasoning problems. For all these evaluations, a higher score is a better score. We chose these benchmarks as they test a variety of reasoning and general knowledge across a wide variety of fields in 0-shot and few-shot settings. Read more [here](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). ### Results | Model | AVG | ARC | Hellaswag | MMLU | Truthful QA | Winogrande | GSM8K | |-------|--------:|------:|----------:|-----:|----------:|----------:|----------:| | [NousResearch/Nous-Puffin-70B](NousResearch/Nous-Puffin-70B) | 64.91 | 67.41 | 87.37 | 69.77 | 46.77 | 83.9 | 34.27 | | [TheBloke/Llama-2-70B-fp16](https://huggingface.co/TheBloke/Llama-2-70B-fp16) | 64.52 | 67.32 | 87.33 | 69.83 | 44.92 | 83.74 | 33.97 | | [NousResearch/Yarn-Mistral-7B-64k](https://huggingface.co/NousResearch/Yarn-Mistral-7b-64k) | 59.63 | 59.9 | 82.51 | 62.96 | 41.86 | 77.27 | 33.28 | | [Qwen1.5-4B-Chat](https://huggingface.co/Qwen/Qwen1.5-4B-Chat) | 46.79 | 43.26 | 69.73 | 55.55 | 44.79 | 64.96 | 2.43 | | [Microsoft/phi-2](https://huggingface.co/microsoft/phi-2) | 61.33 | 61.09 | 75.11 | 58.11 | 44.47 | 74.35 | 54.81 | | [Damysus-2.7B-Chat](https://huggingface.co/prince-canuma/Damysus-2.7B-Chat) (Ours) | 60.49 | 59.81 | 74.52 | 56.33 | **46.74** | **75.06** | 50.64 | ## Technical Specifications ### Compute Infrastructure - Modal Labs #### Hardware - OS: Linux - GPU: A10G #### Libraries - TRL - Transformers - PEFT - Datasets - Accelerate - torch - Wandb - Bitsandbytes - Plotly ## Future work I plan to explore the following tuning setups: - Function calling - DPO ## Citation <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** ```bibtex @misc{Damysus-2.7B-Chat, title={Damysus-2.7B-Chat} , author={Prince Canuma}, year={2024}, } ``` ```bibtex @misc{SlimOrca, title = {SlimOrca: An Open Dataset of GPT-4 Augmented FLAN Reasoning Traces, with Verification}, author = {Wing Lian and Guan Wang and Bleys Goodson and Eugene Pentland and Austin Cook and Chanvichet Vong and "Teknium"}, year = {2023}, publisher = {HuggingFace}, url = {https://https://huggingface.co/Open-Orca/SlimOrca} } ``` ```bibtex @misc{open-llm-leaderboard, author = {Edward Beeching and Clémentine Fourrier and Nathan Habib and Sheon Han and Nathan Lambert and Nazneen Rajani and Omar Sanseviero and Lewis Tunstall and Thomas Wolf}, title = {Open LLM Leaderboard}, year = {2023}, publisher = {Hugging Face}, howpublished = "\url{https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard}" } ```
sunyijia97/falcon-7b-qlora-cstuqa-v6
sunyijia97
2024-02-17T10:06:00Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-02-17T10:05:58Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
FINNUMBER/Yi-Ko-6B-Finch-NQA-EXT-400-epoch8
FINNUMBER
2024-02-17T10:01:05Z
6
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-17T09:25:06Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
hs4jk24erfc/fine_tuned_model_16_02
hs4jk24erfc
2024-02-17T09:40:05Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-02-17T09:39:51Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
FINNUMBER/Yi-Ko-6B-Finch-NQA-ARI-100-epoch16
FINNUMBER
2024-02-17T09:24:59Z
4
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-17T08:27:34Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
alwayssaltyourpasta/my_awesome_mind_model
alwayssaltyourpasta
2024-02-17T09:17:14Z
4
0
transformers
[ "transformers", "tensorboard", "safetensors", "wav2vec2", "audio-classification", "generated_from_trainer", "dataset:minds14", "base_model:facebook/wav2vec2-base", "base_model:finetune:facebook/wav2vec2-base", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
audio-classification
2024-02-16T22:26:08Z
--- license: apache-2.0 base_model: facebook/wav2vec2-base tags: - generated_from_trainer datasets: - minds14 metrics: - accuracy model-index: - name: my_awesome_mind_model results: - task: name: Audio Classification type: audio-classification dataset: name: minds14 type: minds14 config: en-US split: train args: en-US metrics: - name: Accuracy type: accuracy value: 0.05309734513274336 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_awesome_mind_model This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the minds14 dataset. It achieves the following results on the evaluation set: - Loss: 2.6632 - Accuracy: 0.0531 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 0.8 | 3 | 2.6455 | 0.0531 | | No log | 1.87 | 7 | 2.6547 | 0.0708 | | 2.6354 | 2.93 | 11 | 2.6574 | 0.0796 | | 2.6354 | 4.0 | 15 | 2.6594 | 0.0531 | | 2.6354 | 4.8 | 18 | 2.6640 | 0.0442 | | 2.6333 | 5.87 | 22 | 2.6637 | 0.0619 | | 2.6333 | 6.93 | 26 | 2.6632 | 0.0531 | | 2.626 | 8.0 | 30 | 2.6632 | 0.0531 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0 - Datasets 2.17.0 - Tokenizers 0.15.2
FINNUMBER/Yi-Ko-6B-Finch-SA-ESG-100-epoch16
FINNUMBER
2024-02-17T09:10:57Z
3
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-17T07:58:22Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Gordon119/TAT-openai-whisper-large-v2-special-tag-epoch1-total5epoch
Gordon119
2024-02-17T09:09:56Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-02-17T09:09:46Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
sunyijia97/falcon-7b-qlora-cstuqa-v5
sunyijia97
2024-02-17T09:08:41Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-02-17T09:08:38Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
hoanghoavienvo/roberta-base-detect-cheapfake-combined-train-test-2200-2-8
hoanghoavienvo
2024-02-17T09:06:47Z
5
0
transformers
[ "transformers", "tensorboard", "safetensors", "roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/roberta-base", "base_model:finetune:FacebookAI/roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T08:46:59Z
--- license: mit base_model: roberta-base tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: roberta-base-detect-cheapfake-combined-train-test-2200-2-8 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-detect-cheapfake-combined-train-test-2200-2-8 This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4708 - Accuracy: 0.8 - F1: 0.7701 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | No log | 1.0 | 69 | 0.6564 | 0.535 | 0.0971 | | No log | 2.0 | 138 | 0.5171 | 0.725 | 0.6995 | | No log | 3.0 | 207 | 0.4709 | 0.77 | 0.7195 | | No log | 4.0 | 276 | 0.4611 | 0.795 | 0.7630 | | No log | 5.0 | 345 | 0.4708 | 0.8 | 0.7701 | ### Framework versions - Transformers 4.37.0 - Pytorch 2.1.2 - Datasets 2.1.0 - Tokenizers 0.15.1
cheyannelam/lab1_finetuning
cheyannelam
2024-02-17T09:03:29Z
3
0
transformers
[ "transformers", "safetensors", "marian", "text2text-generation", "generated_from_trainer", "dataset:kde4", "base_model:Helsinki-NLP/opus-mt-en-fr", "base_model:finetune:Helsinki-NLP/opus-mt-en-fr", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-02-15T21:10:44Z
--- license: apache-2.0 base_model: Helsinki-NLP/opus-mt-en-fr tags: - generated_from_trainer datasets: - kde4 metrics: - bleu model-index: - name: marian-finetuned-kde4-en-to-fr results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: kde4 type: kde4 config: en-fr split: train args: en-fr metrics: - name: Bleu type: bleu value: 13.70591658813882 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # marian-finetuned-kde4-en-to-fr This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset. It achieves the following results on the evaluation set: - Loss: 3.4909 - Bleu: 13.7059 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
badokorach/distilbert-base-cased-distilled-agric-170224
badokorach
2024-02-17T09:01:07Z
5
0
transformers
[ "transformers", "tf", "distilbert", "question-answering", "generated_from_keras_callback", "base_model:badokorach/distilbert-base-cased-distilled-agric-060124_1", "base_model:finetune:badokorach/distilbert-base-cased-distilled-agric-060124_1", "license:apache-2.0", "endpoints_compatible", "region:us" ]
question-answering
2024-02-17T08:39:39Z
--- license: apache-2.0 base_model: badokorach/distilbert-base-cased-distilled-agric-060124_1 tags: - generated_from_keras_callback model-index: - name: badokorach/distilbert-base-cased-distilled-agric-170224 results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # badokorach/distilbert-base-cased-distilled-agric-170224 This model is a fine-tuned version of [badokorach/distilbert-base-cased-distilled-agric-060124_1](https://huggingface.co/badokorach/distilbert-base-cased-distilled-agric-060124_1) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.0008 - Validation Loss: 0.0 - Epoch: 4 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'inner_optimizer': {'module': 'keras.optimizers', 'class_name': 'Adam', 'config': {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': True, 'is_legacy_optimizer': False, 'learning_rate': {'module': None, 'class_name': 'CustomLearningRateScheduler', 'config': {'initial_learning_rate': 3e-05, 'decay_steps': 2736, 'warmup_steps': 304, 'end_learning_rate': 1e-05}, 'registered_name': 'CustomLearningRateScheduler'}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}, 'registered_name': None}, 'dynamic': True, 'initial_scale': 32768.0, 'dynamic_growth_steps': 2000} - training_precision: mixed_float16 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 0.4220 | 0.0 | 0 | | 0.0117 | 0.0 | 1 | | 0.0066 | 0.0 | 2 | | 0.0055 | 0.0 | 3 | | 0.0008 | 0.0 | 4 | ### Framework versions - Transformers 4.35.2 - TensorFlow 2.15.0 - Datasets 2.17.0 - Tokenizers 0.15.2
Azma-AI/azma-starling-LM-7B-alpha-agent-v1
Azma-AI
2024-02-17T08:56:31Z
6
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-17T08:52:06Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
likhith231/opus-mt-en-ro-finetuned-en-to-ro
likhith231
2024-02-17T08:46:47Z
7
0
transformers
[ "transformers", "tensorboard", "safetensors", "marian", "text2text-generation", "generated_from_trainer", "base_model:Helsinki-NLP/opus-mt-en-ro", "base_model:finetune:Helsinki-NLP/opus-mt-en-ro", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-02-17T07:15:56Z
--- license: apache-2.0 base_model: Helsinki-NLP/opus-mt-en-ro tags: - generated_from_trainer metrics: - bleu model-index: - name: opus-mt-en-ro-finetuned-en-to-ro results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # opus-mt-en-ro-finetuned-en-to-ro This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-ro](https://huggingface.co/Helsinki-NLP/opus-mt-en-ro) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4095 - Bleu: 26.3441 - Gen Len: 34.093 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:| | No log | 1.0 | 313 | 0.4031 | 26.665 | 33.886 | | 0.1065 | 2.0 | 626 | 0.4074 | 26.3571 | 34.178 | | 0.1065 | 3.0 | 939 | 0.4095 | 26.3441 | 34.093 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
hoanghoavienvo/roberta-base-detect-cheapfake-combined-train-test-contradict-2-8
hoanghoavienvo
2024-02-17T08:43:48Z
3
0
transformers
[ "transformers", "tensorboard", "safetensors", "roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/roberta-base", "base_model:finetune:FacebookAI/roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T08:20:14Z
--- license: mit base_model: roberta-base tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: roberta-base-detect-cheapfake-combined-train-test-contradict-2-8 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-detect-cheapfake-combined-train-test-contradict-2-8 This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.5226 - Accuracy: 0.835 - F1: 0.8156 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | No log | 1.0 | 163 | 0.7064 | 0.64 | 0.5385 | | No log | 2.0 | 326 | 0.5252 | 0.765 | 0.7662 | | No log | 3.0 | 489 | 0.4988 | 0.82 | 0.8269 | | 0.1701 | 4.0 | 652 | 0.6552 | 0.77 | 0.7125 | | 0.1701 | 5.0 | 815 | 0.5226 | 0.835 | 0.8156 | ### Framework versions - Transformers 4.37.0 - Pytorch 2.1.2 - Datasets 2.1.0 - Tokenizers 0.15.1
callmesan/vakyansh-wav2vec2-tamil-tam-250-audio-abuse-feature
callmesan
2024-02-17T08:40:32Z
2
0
transformers
[ "transformers", "pytorch", "wav2vec2", "audio-classification", "generated_from_trainer", "base_model:Harveenchadha/vakyansh-wav2vec2-tamil-tam-250", "base_model:finetune:Harveenchadha/vakyansh-wav2vec2-tamil-tam-250", "license:mit", "endpoints_compatible", "region:us" ]
audio-classification
2024-02-17T08:20:27Z
--- license: mit base_model: Harveenchadha/vakyansh-wav2vec2-tamil-tam-250 tags: - generated_from_trainer metrics: - accuracy model-index: - name: vakyansh-wav2vec2-tamil-tam-250-audio-abuse-feature results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vakyansh-wav2vec2-tamil-tam-250-audio-abuse-feature This model is a fine-tuned version of [Harveenchadha/vakyansh-wav2vec2-tamil-tam-250](https://huggingface.co/Harveenchadha/vakyansh-wav2vec2-tamil-tam-250) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6061 - Accuracy: 0.7412 - Macro F1-score: 0.6531 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 50 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Macro F1-score | |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------------:| | 6.7458 | 0.77 | 10 | 6.7472 | 0.0 | 0.0 | | 6.7056 | 1.54 | 20 | 6.6488 | 0.0 | 0.0 | | 6.6158 | 2.31 | 30 | 6.5180 | 0.6307 | 0.0917 | | 6.4651 | 3.08 | 40 | 6.2887 | 0.7143 | 0.4167 | | 6.2508 | 3.85 | 50 | 5.9094 | 0.7197 | 0.4185 | | 5.8959 | 4.62 | 60 | 5.5362 | 0.7197 | 0.4185 | | 5.6179 | 5.38 | 70 | 5.2347 | 0.7197 | 0.4185 | | 5.3048 | 6.15 | 80 | 4.9823 | 0.7197 | 0.4185 | | 5.0858 | 6.92 | 90 | 4.7555 | 0.7197 | 0.4185 | | 4.9195 | 7.69 | 100 | 4.5424 | 0.7197 | 0.4185 | | 4.6747 | 8.46 | 110 | 4.3265 | 0.7197 | 0.4185 | | 4.5861 | 9.23 | 120 | 4.1193 | 0.7197 | 0.4185 | | 4.3397 | 10.0 | 130 | 3.9070 | 0.7197 | 0.4185 | | 4.0926 | 10.77 | 140 | 3.6954 | 0.7197 | 0.4185 | | 3.8859 | 11.54 | 150 | 3.4822 | 0.7197 | 0.4185 | | 3.7254 | 12.31 | 160 | 3.2711 | 0.7197 | 0.4185 | | 3.5303 | 13.08 | 170 | 3.0599 | 0.7197 | 0.4185 | | 3.2531 | 13.85 | 180 | 2.8502 | 0.7197 | 0.4185 | | 3.0184 | 14.62 | 190 | 2.6448 | 0.7197 | 0.4185 | | 3.0006 | 15.38 | 200 | 2.4472 | 0.7197 | 0.4185 | | 2.6674 | 16.15 | 210 | 2.2526 | 0.7197 | 0.4185 | | 2.4455 | 16.92 | 220 | 2.0649 | 0.7197 | 0.4185 | | 2.2702 | 17.69 | 230 | 1.8883 | 0.7197 | 0.4185 | | 2.0536 | 18.46 | 240 | 1.7233 | 0.7197 | 0.4185 | | 2.0643 | 19.23 | 250 | 1.5730 | 0.7197 | 0.4185 | | 1.8006 | 20.0 | 260 | 1.4368 | 0.7197 | 0.4185 | | 1.6975 | 20.77 | 270 | 1.3112 | 0.7197 | 0.4185 | | 1.4407 | 21.54 | 280 | 1.2015 | 0.7197 | 0.4185 | | 1.2971 | 22.31 | 290 | 1.1050 | 0.7197 | 0.4185 | | 1.3202 | 23.08 | 300 | 1.0219 | 0.7197 | 0.4185 | | 1.1292 | 23.85 | 310 | 0.9490 | 0.7197 | 0.4185 | | 1.1055 | 24.62 | 320 | 0.8879 | 0.7197 | 0.4185 | | 0.9817 | 25.38 | 330 | 0.8366 | 0.7197 | 0.4185 | | 0.9296 | 26.15 | 340 | 0.7906 | 0.7197 | 0.4185 | | 0.8306 | 26.92 | 350 | 0.7506 | 0.7197 | 0.4185 | | 0.8303 | 27.69 | 360 | 0.7171 | 0.7197 | 0.4185 | | 0.8421 | 28.46 | 370 | 0.6953 | 0.7197 | 0.4185 | | 0.7964 | 29.23 | 380 | 0.6650 | 0.7197 | 0.4185 | | 0.7528 | 30.0 | 390 | 0.6470 | 0.7197 | 0.4185 | | 0.7305 | 30.77 | 400 | 0.6345 | 0.7197 | 0.4185 | | 0.6702 | 31.54 | 410 | 0.6163 | 0.7385 | 0.4937 | | 0.6416 | 32.31 | 420 | 0.6118 | 0.7547 | 0.5507 | | 0.608 | 33.08 | 430 | 0.6086 | 0.7547 | 0.5507 | | 0.6659 | 33.85 | 440 | 0.5981 | 0.7574 | 0.5949 | | 0.5839 | 34.62 | 450 | 0.6068 | 0.7547 | 0.6570 | | 0.6167 | 35.38 | 460 | 0.5894 | 0.7763 | 0.6479 | | 0.5991 | 36.15 | 470 | 0.5947 | 0.7412 | 0.6531 | | 0.5839 | 36.92 | 480 | 0.5938 | 0.7574 | 0.6771 | | 0.5533 | 37.69 | 490 | 0.5922 | 0.7520 | 0.6399 | | 0.4998 | 38.46 | 500 | 0.6203 | 0.7358 | 0.6625 | | 0.5508 | 39.23 | 510 | 0.5865 | 0.7493 | 0.6278 | | 0.5159 | 40.0 | 520 | 0.5963 | 0.7385 | 0.6670 | | 0.5344 | 40.77 | 530 | 0.5946 | 0.7439 | 0.6420 | | 0.5039 | 41.54 | 540 | 0.5979 | 0.7466 | 0.6526 | | 0.5456 | 42.31 | 550 | 0.5999 | 0.7358 | 0.6707 | | 0.4822 | 43.08 | 560 | 0.5845 | 0.7493 | 0.6437 | | 0.4864 | 43.85 | 570 | 0.6035 | 0.7439 | 0.6779 | | 0.4623 | 44.62 | 580 | 0.5961 | 0.7520 | 0.6519 | | 0.475 | 45.38 | 590 | 0.6066 | 0.7439 | 0.6651 | | 0.4887 | 46.15 | 600 | 0.6014 | 0.7466 | 0.6603 | | 0.506 | 46.92 | 610 | 0.6012 | 0.7412 | 0.6604 | | 0.5296 | 47.69 | 620 | 0.5986 | 0.7439 | 0.6503 | | 0.5255 | 48.46 | 630 | 0.6003 | 0.7439 | 0.6503 | | 0.4667 | 49.23 | 640 | 0.6038 | 0.7466 | 0.6553 | | 0.4334 | 50.0 | 650 | 0.6061 | 0.7412 | 0.6531 | ### Framework versions - Transformers 4.33.0 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3
mesolitica/malaysian-tinyllama-1.1b-siglip-large-384-vision
mesolitica
2024-02-17T08:39:21Z
10
0
transformers
[ "transformers", "safetensors", "mm_llms", "feature-extraction", "custom_code", "region:us" ]
feature-extraction
2024-02-12T02:11:22Z
--- library_name: transformers tags: [] --- # Malaysian TinyLlama + siglip-large-patch16-384 WanDB https://wandb.ai/huseinzol05/vision-tinyllama?workspace=user-huseinzol05 ## how-to ```python from modeling_vision import MM_LLMs, MM_LLMs_Config from transformers import AutoTokenizer, AutoProcessor from PIL import Image import requests model = MM_LLMs.from_pretrained( 'mesolitica/malaysian-tinyllama-1.1b-siglip-large-384-vision', flash_attention = True, dtype = torch.bfloat16, torch_dtype = torch.bfloat16 ) _ = model.cuda() image_processor = AutoProcessor.from_pretrained('google/siglip-large-patch16-384') tokenizer = AutoTokenizer.from_pretrained('mesolitica/malaysian-tinyllama-1.1b-siglip-large-384-vision') def prepare_dataset(messages, images: List[str] = None): if images is not None: images = [Image.open(f).convert('RGB') for f in images] image_output = image_processor(images=images, return_tensors='pt')['pixel_values'] else: image_output = None prompt = tokenizer.apply_chat_template(messages, tokenize = False) outputs = tokenizer( prompt, return_tensors='pt', return_overflowing_tokens=False, return_length=False) outputs['images'] = image_output outputs['image_index'] = torch.tensor([0] * len(outputs['images'])) outputs['image_starts'] = torch.tensor([tokenizer.convert_tokens_to_ids('<image>')] * len(outputs['images'])) return outputs with open('Persian-cat-breed.jpg', 'wb') as fopen: fopen.write(requests.get('https://cdn.beautifulnara.net/wp-content/uploads/2017/12/10201620/Persian-cat-breed.jpg').content) with open('nasi-goreng-1-23.jpg', 'wb') as fopen: fopen.write(requests.get('https://www.jocooks.com/wp-content/uploads/2023/09/nasi-goreng-1-23.jpg').content) messages = [ {'role': 'user', 'content': '<image> </image> ini gambar apa'}, ] outputs = prepare_dataset(messages, images = ['Persian-cat-breed.jpg']) outputs['images'] = outputs['images'].type(model.dtype) for k in outputs.keys(): if outputs[k] is not None: outputs[k] = outputs[k].cuda() with torch.no_grad(): model_inputs = model.prepare_inputs_for_generation(**outputs) r = model_inputs.pop('input_ids', None) generate_kwargs = dict( model_inputs, max_new_tokens=300, top_p=0.95, top_k=50, temperature=0.1, do_sample=True, num_beams=1, ) r = model.llm.generate(**generate_kwargs) print(tokenizer.decode(r[0])) ``` ``` <s>Imej itu menunjukkan seekor kucing putih yang comel duduk di atas sofa hitam.</s> ``` ```python messages = [ {'role': 'user', 'content': '<image> </image> <image> </image> apa kaitan 2 gambar ni'}, ] outputs = prepare_dataset(messages, images = ['Persian-cat-breed.jpg', 'nasi-goreng-1-23.jpg']) outputs['images'] = outputs['images'].type(model.dtype) for k in outputs.keys(): if outputs[k] is not None: outputs[k] = outputs[k].cuda() with torch.no_grad(): model_inputs = model.prepare_inputs_for_generation(**outputs) r = model_inputs.pop('input_ids', None) generate_kwargs = dict( model_inputs, max_new_tokens=300, top_p=0.95, top_k=50, temperature=0.1, do_sample=True, num_beams=1, ) r = model.llm.generate(**generate_kwargs) print(tokenizer.decode(r[0])) ``` ``` <s>Tiada hubungan yang jelas antara gambar 1 (anak kucing putih duduk di atas sofa) dan gambar 2 (foto penutup mangkuk mi telur dengan nasi dan cili). Gambar pertama ialah imej haiwan, manakala gambar kedua ialah imej makanan. Mereka tergolong dalam kategori yang berbeza dan tidak mempunyai hubungan antara satu sama lain.</s> ```
mesolitica/malaysian-Qwen1.5-0.5B-siglip-base-384-vision
mesolitica
2024-02-17T08:38:07Z
4
1
transformers
[ "transformers", "pytorch", "mm_llms", "feature-extraction", "custom_code", "region:us" ]
feature-extraction
2024-02-12T04:41:11Z
--- library_name: transformers tags: [] --- # Malaysian Qwen1.5-0.5B + siglip-base-patch16-384 WanDB https://wandb.ai/huseinzol05/vision-qwen0.5?workspace=user-huseinzol05 ## how-to ```python from modeling_vision import MM_LLMs, MM_LLMs_Config from transformers import AutoTokenizer, AutoProcessor from PIL import Image import requests model = MM_LLMs.from_pretrained( 'mesolitica/malaysian-Qwen1.5-0.5B-siglip-base-384-vision', flash_attention = True, dtype = torch.bfloat16, torch_dtype = torch.bfloat16 ) _ = model.cuda() image_processor = AutoProcessor.from_pretrained('google/siglip-base-patch16-384') tokenizer = AutoTokenizer.from_pretrained('mesolitica/malaysian-Qwen1.5-0.5B-siglip-base-384-vision') model.llm.generation_config.eos_token_id = tokenizer.eos_token_id def prepare_dataset(messages, images: List[str] = None): if images is not None: images = [Image.open(f).convert('RGB') for f in images] image_output = image_processor(images=images, return_tensors='pt')['pixel_values'] else: image_output = None prompt = tokenizer.apply_chat_template(messages, tokenize = False) outputs = tokenizer( prompt, return_tensors='pt', return_overflowing_tokens=False, return_length=False) outputs['images'] = image_output outputs['image_index'] = torch.tensor([0] * len(outputs['images'])) outputs['image_starts'] = torch.tensor([tokenizer.convert_tokens_to_ids('<image>')] * len(outputs['images'])) return outputs with open('Persian-cat-breed.jpg', 'wb') as fopen: fopen.write(requests.get('https://cdn.beautifulnara.net/wp-content/uploads/2017/12/10201620/Persian-cat-breed.jpg').content) with open('nasi-goreng-1-23.jpg', 'wb') as fopen: fopen.write(requests.get('https://www.jocooks.com/wp-content/uploads/2023/09/nasi-goreng-1-23.jpg').content) messages = [ {'role': 'user', 'content': '<image> </image> ini gambar apa'}, ] outputs = prepare_dataset(messages, images = ['Persian-cat-breed.jpg']) outputs['images'] = outputs['images'].type(model.dtype) for k in outputs.keys(): if outputs[k] is not None: outputs[k] = outputs[k].cuda() with torch.no_grad(): model_inputs = model.prepare_inputs_for_generation(**outputs) r = model_inputs.pop('input_ids', None) generate_kwargs = dict( model_inputs, max_new_tokens=300, top_p=0.95, top_k=50, temperature=0.1, do_sample=True, num_beams=1, ) r = model.llm.generate(**generate_kwargs) print(tokenizer.decode(r[0])) ``` ``` <|endoftext|><|im_start|>assistant Ini adalah gambar kucing putih yang duduk di atas sofa hitam.<|im_end|> ``` ```python messages = [ {'role': 'user', 'content': '<image> </image> <image> </image> apa kaitan 2 gambar ni'}, ] outputs = prepare_dataset(messages, images = ['Persian-cat-breed.jpg', 'nasi-goreng-1-23.jpg']) outputs['images'] = outputs['images'].type(model.dtype) for k in outputs.keys(): if outputs[k] is not None: outputs[k] = outputs[k].cuda() with torch.no_grad(): model_inputs = model.prepare_inputs_for_generation(**outputs) r = model_inputs.pop('input_ids', None) generate_kwargs = dict( model_inputs, max_new_tokens=300, top_p=0.95, top_k=50, temperature=0.1, do_sample=True, num_beams=1, ) r = model.llm.generate(**generate_kwargs) print(tokenizer.decode(r[0])) ``` ``` <|endoftext|><|im_start|>assistant Tiada hubungan langsung antara gambar 1 dan gambar 2. Gambar 1 ialah imej kucing putih dengan bulu putih, manakala gambar 2 ialah gambar mangkuk makan tengah hari kacang hitam dan lobak merah yang dicincang, dengan garpu diletakkan di sebelahnya. Kedua-duanya tidak berkaitan dari segi kandungan.<|im_end|> ```
migueldeguzmandev/Phi-1.5-RLLMv3-8
migueldeguzmandev
2024-02-17T08:28:01Z
4
0
transformers
[ "transformers", "pytorch", "phi", "text-generation", "custom_code", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-21T09:06:14Z
Companion Post: [Research Log, RLLMv3 (GPT2-XL, Phi-1.5 and Falcon-RW-1B)](https://www.lesswrong.com/posts/EiEhYmYsvYCRgCemH/research-log-rllmv3-gpt2-xl-phi-1-5-and-falcon-rw-1b?utm_campaign=post_share&utm_source=link) Main post: [BetterDAN, AI Machiavelli & Oppo Jailbreaks vs. SOTA models & GPT2XL_RLLMv3](https://www.lesswrong.com/posts/vZ5fM6FtriyyKbwi9/betterdan-ai-machiavelli-and-oppo-jailbreaks-vs-sota-models?utm_campaign=post_share&utm_source=link) Related post: [Coherence (and Response Time) Test](https://docs.google.com/document/d/1D235vN2KwsLIUKCySpKJoDLV7qwYcU-LSSDpFCbMljs/edit?usp=sharing)
migueldeguzmandev/Phi-1.5-RLLMv3-3
migueldeguzmandev
2024-02-17T08:26:54Z
4
0
transformers
[ "transformers", "pytorch", "phi", "text-generation", "custom_code", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-21T08:08:12Z
Companion Post: [Research Log, RLLMv3 (GPT2-XL, Phi-1.5 and Falcon-RW-1B)](https://www.lesswrong.com/posts/EiEhYmYsvYCRgCemH/research-log-rllmv3-gpt2-xl-phi-1-5-and-falcon-rw-1b?utm_campaign=post_share&utm_source=link) Main post: [BetterDAN, AI Machiavelli & Oppo Jailbreaks vs. SOTA models & GPT2XL_RLLMv3](https://www.lesswrong.com/posts/vZ5fM6FtriyyKbwi9/betterdan-ai-machiavelli-and-oppo-jailbreaks-vs-sota-models?utm_campaign=post_share&utm_source=link) Related post: [Coherence (and Response Time) Test](https://docs.google.com/document/d/1D235vN2KwsLIUKCySpKJoDLV7qwYcU-LSSDpFCbMljs/edit?usp=sharing)
SecondTheFirst/poca-SoccerTwos
SecondTheFirst
2024-02-17T08:22:37Z
14
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "SoccerTwos", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SoccerTwos", "region:us" ]
reinforcement-learning
2024-02-17T08:21:56Z
--- library_name: ml-agents tags: - SoccerTwos - deep-reinforcement-learning - reinforcement-learning - ML-Agents-SoccerTwos --- # **poca** Agent playing **SoccerTwos** This is a trained model of a **poca** agent playing **SoccerTwos** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: SecondTheFirst/poca-SoccerTwos 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
longcule123/lora_model172
longcule123
2024-02-17T07:55:53Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-02-17T07:55:49Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Falconsai/text_summarization
Falconsai
2024-02-17T07:55:14Z
68,083
212
transformers
[ "transformers", "pytorch", "coreml", "onnx", "safetensors", "t5", "text2text-generation", "summarization", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
summarization
2023-10-21T00:53:53Z
--- license: apache-2.0 language: - en pipeline_tag: summarization widget: - text: >- Hugging Face: Revolutionizing Natural Language Processing Introduction In the rapidly evolving field of Natural Language Processing (NLP), Hugging Face has emerged as a prominent and innovative force. This article will explore the story and significance of Hugging Face, a company that has made remarkable contributions to NLP and AI as a whole. From its inception to its role in democratizing AI, Hugging Face has left an indelible mark on the industry. The Birth of Hugging Face Hugging Face was founded in 2016 by Clément Delangue, Julien Chaumond, and Thomas Wolf. The name Hugging Face was chosen to reflect the company's mission of making AI models more accessible and friendly to humans, much like a comforting hug. Initially, they began as a chatbot company but later shifted their focus to NLP, driven by their belief in the transformative potential of this technology. Transformative Innovations Hugging Face is best known for its open-source contributions, particularly the Transformers library. This library has become the de facto standard for NLP and enables researchers, developers, and organizations to easily access and utilize state-of-the-art pre-trained language models, such as BERT, GPT-3, and more. These models have countless applications, from chatbots and virtual assistants to language translation and sentiment analysis. example_title: Summarization Example 1 --- # Model Card: Fine-Tuned T5 Small for Text Summarization ## Model Description The **Fine-Tuned T5 Small** is a variant of the T5 transformer model, designed for the task of text summarization. It is adapted and fine-tuned to generate concise and coherent summaries of input text. The model, named "t5-small," is pre-trained on a diverse corpus of text data, enabling it to capture essential information and generate meaningful summaries. Fine-tuning is conducted with careful attention to hyperparameter settings, including batch size and learning rate, to ensure optimal performance for text summarization. During the fine-tuning process, a batch size of 8 is chosen for efficient computation and learning. Additionally, a learning rate of 2e-5 is selected to balance convergence speed and model optimization. This approach guarantees not only rapid learning but also continuous refinement during training. The fine-tuning dataset consists of a variety of documents and their corresponding human-generated summaries. This diverse dataset allows the model to learn the art of creating summaries that capture the most important information while maintaining coherence and fluency. The goal of this meticulous training process is to equip the model with the ability to generate high-quality text summaries, making it valuable for a wide range of applications involving document summarization and content condensation. ## Intended Uses & Limitations ### Intended Uses - **Text Summarization**: The primary intended use of this model is to generate concise and coherent text summaries. It is well-suited for applications that involve summarizing lengthy documents, news articles, and textual content. ### How to Use To use this model for text summarization, you can follow these steps: ```python from transformers import pipeline summarizer = pipeline("summarization", model="Falconsai/text_summarization") ARTICLE = """ Hugging Face: Revolutionizing Natural Language Processing Introduction In the rapidly evolving field of Natural Language Processing (NLP), Hugging Face has emerged as a prominent and innovative force. This article will explore the story and significance of Hugging Face, a company that has made remarkable contributions to NLP and AI as a whole. From its inception to its role in democratizing AI, Hugging Face has left an indelible mark on the industry. The Birth of Hugging Face Hugging Face was founded in 2016 by Clément Delangue, Julien Chaumond, and Thomas Wolf. The name "Hugging Face" was chosen to reflect the company's mission of making AI models more accessible and friendly to humans, much like a comforting hug. Initially, they began as a chatbot company but later shifted their focus to NLP, driven by their belief in the transformative potential of this technology. Transformative Innovations Hugging Face is best known for its open-source contributions, particularly the "Transformers" library. This library has become the de facto standard for NLP and enables researchers, developers, and organizations to easily access and utilize state-of-the-art pre-trained language models, such as BERT, GPT-3, and more. These models have countless applications, from chatbots and virtual assistants to language translation and sentiment analysis. Key Contributions: 1. **Transformers Library:** The Transformers library provides a unified interface for more than 50 pre-trained models, simplifying the development of NLP applications. It allows users to fine-tune these models for specific tasks, making it accessible to a wider audience. 2. **Model Hub:** Hugging Face's Model Hub is a treasure trove of pre-trained models, making it simple for anyone to access, experiment with, and fine-tune models. Researchers and developers around the world can collaborate and share their models through this platform. 3. **Hugging Face Transformers Community:** Hugging Face has fostered a vibrant online community where developers, researchers, and AI enthusiasts can share their knowledge, code, and insights. This collaborative spirit has accelerated the growth of NLP. Democratizing AI Hugging Face's most significant impact has been the democratization of AI and NLP. Their commitment to open-source development has made powerful AI models accessible to individuals, startups, and established organizations. This approach contrasts with the traditional proprietary AI model market, which often limits access to those with substantial resources. By providing open-source models and tools, Hugging Face has empowered a diverse array of users to innovate and create their own NLP applications. This shift has fostered inclusivity, allowing a broader range of voices to contribute to AI research and development. Industry Adoption The success and impact of Hugging Face are evident in its widespread adoption. Numerous companies and institutions, from startups to tech giants, leverage Hugging Face's technology for their AI applications. This includes industries as varied as healthcare, finance, and entertainment, showcasing the versatility of NLP and Hugging Face's contributions. Future Directions Hugging Face's journey is far from over. As of my last knowledge update in September 2021, the company was actively pursuing research into ethical AI, bias reduction in models, and more. Given their track record of innovation and commitment to the AI community, it is likely that they will continue to lead in ethical AI development and promote responsible use of NLP technologies. Conclusion Hugging Face's story is one of transformation, collaboration, and empowerment. Their open-source contributions have reshaped the NLP landscape and democratized access to AI. As they continue to push the boundaries of AI research, we can expect Hugging Face to remain at the forefront of innovation, contributing to a more inclusive and ethical AI future. Their journey reminds us that the power of open-source collaboration can lead to groundbreaking advancements in technology and bring AI within the reach of many. """ print(summarizer(ARTICLE, max_length=1000, min_length=30, do_sample=False)) >>> [{'summary_text': 'Hugging Face has emerged as a prominent and innovative force in NLP . From its inception to its role in democratizing AI, the company has left an indelible mark on the industry . The name "Hugging Face" was chosen to reflect the company\'s mission of making AI models more accessible and friendly to humans .'}] ``` Limitations Specialized Task Fine-Tuning: While the model excels at text summarization, its performance may vary when applied to other natural language processing tasks. Users interested in employing this model for different tasks should explore fine-tuned versions available in the model hub for optimal results. Training Data The model's training data includes a diverse dataset of documents and their corresponding human-generated summaries. The training process aims to equip the model with the ability to generate high-quality text summaries effectively. Training Stats - Evaluation Loss: 0.012345678901234567 - Evaluation Rouge Score: 0.95 (F1) - Evaluation Runtime: 2.3456 - Evaluation Samples per Second: 1234.56 - Evaluation Steps per Second: 45.678 Responsible Usage It is essential to use this model responsibly and ethically, adhering to content guidelines and applicable regulations when implementing it in real-world applications, particularly those involving potentially sensitive content. References Hugging Face Model Hub T5 Paper Disclaimer: The model's performance may be influenced by the quality and representativeness of the data it was fine-tuned on. Users are encouraged to assess the model's suitability for their specific applications and datasets.
callmesan/vakyansh-wav2vec2-odia-orm-100-audio-abuse-feature
callmesan
2024-02-17T07:41:03Z
2
0
transformers
[ "transformers", "pytorch", "wav2vec2", "audio-classification", "generated_from_trainer", "base_model:Harveenchadha/vakyansh-wav2vec2-odia-orm-100", "base_model:finetune:Harveenchadha/vakyansh-wav2vec2-odia-orm-100", "endpoints_compatible", "region:us" ]
audio-classification
2024-02-17T07:22:26Z
--- base_model: Harveenchadha/vakyansh-wav2vec2-odia-orm-100 tags: - generated_from_trainer metrics: - accuracy model-index: - name: vakyansh-wav2vec2-odia-orm-100-audio-abuse-feature results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vakyansh-wav2vec2-odia-orm-100-audio-abuse-feature This model is a fine-tuned version of [Harveenchadha/vakyansh-wav2vec2-odia-orm-100](https://huggingface.co/Harveenchadha/vakyansh-wav2vec2-odia-orm-100) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7299 - Accuracy: 0.7014 - Macro F1-score: 0.6792 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 50 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Macro F1-score | |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------------:| | 6.7078 | 0.78 | 10 | 6.6948 | 0.0 | 0.0 | | 6.6539 | 1.57 | 20 | 6.5580 | 0.2 | 0.0342 | | 6.5111 | 2.35 | 30 | 6.3377 | 0.5726 | 0.3641 | | 6.268 | 3.14 | 40 | 6.0361 | 0.5726 | 0.3641 | | 6.0748 | 3.92 | 50 | 5.7417 | 0.5726 | 0.3641 | | 5.8205 | 4.71 | 60 | 5.4985 | 0.5726 | 0.3641 | | 5.6051 | 5.49 | 70 | 5.2743 | 0.5726 | 0.3641 | | 5.3589 | 6.27 | 80 | 5.0823 | 0.5726 | 0.3641 | | 5.2019 | 7.06 | 90 | 4.8953 | 0.5726 | 0.3641 | | 5.0528 | 7.84 | 100 | 4.7077 | 0.5726 | 0.3641 | | 4.868 | 8.63 | 110 | 4.5244 | 0.5726 | 0.3641 | | 4.7081 | 9.41 | 120 | 4.3347 | 0.5726 | 0.3641 | | 4.437 | 10.2 | 130 | 4.1455 | 0.5726 | 0.3641 | | 4.3225 | 10.98 | 140 | 3.9551 | 0.5726 | 0.3641 | | 4.0945 | 11.76 | 150 | 3.7694 | 0.5726 | 0.3641 | | 4.014 | 12.55 | 160 | 3.5710 | 0.5726 | 0.3641 | | 3.8491 | 13.33 | 170 | 3.3814 | 0.5726 | 0.3641 | | 3.4724 | 14.12 | 180 | 3.1873 | 0.5726 | 0.3641 | | 3.2728 | 14.9 | 190 | 2.9999 | 0.5726 | 0.3641 | | 3.1948 | 15.69 | 200 | 2.8224 | 0.5726 | 0.3641 | | 2.9968 | 16.47 | 210 | 2.6368 | 0.5726 | 0.3641 | | 2.6739 | 17.25 | 220 | 2.4462 | 0.5726 | 0.3641 | | 2.561 | 18.04 | 230 | 2.2871 | 0.5726 | 0.3641 | | 2.5101 | 18.82 | 240 | 2.1260 | 0.5726 | 0.3641 | | 2.3307 | 19.61 | 250 | 1.9620 | 0.5726 | 0.3641 | | 2.1022 | 20.39 | 260 | 1.8260 | 0.5726 | 0.3641 | | 1.9909 | 21.18 | 270 | 1.6933 | 0.5726 | 0.3641 | | 1.766 | 21.96 | 280 | 1.5644 | 0.5726 | 0.3641 | | 1.7143 | 22.75 | 290 | 1.4669 | 0.5726 | 0.3641 | | 1.5073 | 23.53 | 300 | 1.3482 | 0.5726 | 0.3641 | | 1.6055 | 24.31 | 310 | 1.2643 | 0.5726 | 0.3641 | | 1.321 | 25.1 | 320 | 1.1930 | 0.5726 | 0.3641 | | 1.2165 | 25.88 | 330 | 1.1128 | 0.5726 | 0.3641 | | 1.1484 | 26.67 | 340 | 1.0493 | 0.6712 | 0.6033 | | 1.1413 | 27.45 | 350 | 0.9925 | 0.7096 | 0.6737 | | 1.0462 | 28.24 | 360 | 0.9471 | 0.6877 | 0.6190 | | 0.9667 | 29.02 | 370 | 0.9209 | 0.7123 | 0.6869 | | 0.9918 | 29.8 | 380 | 0.8892 | 0.7205 | 0.6953 | | 0.9112 | 30.59 | 390 | 0.8414 | 0.7123 | 0.6705 | | 0.8666 | 31.37 | 400 | 0.8291 | 0.7123 | 0.6836 | | 0.8096 | 32.16 | 410 | 0.8284 | 0.6959 | 0.6501 | | 0.7987 | 32.94 | 420 | 0.7729 | 0.7425 | 0.7270 | | 0.7529 | 33.73 | 430 | 0.7542 | 0.7260 | 0.7023 | | 0.7605 | 34.51 | 440 | 0.7535 | 0.7260 | 0.7043 | | 0.7011 | 35.29 | 450 | 0.7882 | 0.6959 | 0.6891 | | 0.6868 | 36.08 | 460 | 0.7378 | 0.7260 | 0.7013 | | 0.6858 | 36.86 | 470 | 0.7518 | 0.7096 | 0.6865 | | 0.7546 | 37.65 | 480 | 0.7163 | 0.7342 | 0.7108 | | 0.6717 | 38.43 | 490 | 0.7158 | 0.7397 | 0.7158 | | 0.7048 | 39.22 | 500 | 0.7755 | 0.6575 | 0.6487 | | 0.6767 | 40.0 | 510 | 0.7469 | 0.7068 | 0.6798 | | 0.6621 | 40.78 | 520 | 0.7166 | 0.7205 | 0.7020 | | 0.6639 | 41.57 | 530 | 0.7143 | 0.7151 | 0.6934 | | 0.5988 | 42.35 | 540 | 0.7547 | 0.6767 | 0.6661 | | 0.6179 | 43.14 | 550 | 0.7394 | 0.7014 | 0.6820 | | 0.7033 | 43.92 | 560 | 0.7312 | 0.6986 | 0.6757 | | 0.6076 | 44.71 | 570 | 0.7331 | 0.6904 | 0.6674 | | 0.602 | 45.49 | 580 | 0.7341 | 0.6932 | 0.6718 | | 0.545 | 46.27 | 590 | 0.7363 | 0.6932 | 0.6738 | | 0.5881 | 47.06 | 600 | 0.7299 | 0.7014 | 0.6792 | ### Framework versions - Transformers 4.33.0 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3
Deadwalker0/maverick-34b-qlora
Deadwalker0
2024-02-17T07:31:41Z
0
0
peft
[ "peft", "tensorboard", "safetensors", "llama", "generated_from_trainer", "base_model:codellama/CodeLlama-34b-hf", "base_model:adapter:codellama/CodeLlama-34b-hf", "license:llama2", "4-bit", "bitsandbytes", "region:us" ]
null
2024-02-17T07:24:27Z
--- license: llama2 library_name: peft tags: - generated_from_trainer base_model: codellama/CodeLlama-34b-hf model-index: - name: maverick34b results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.0` ```yaml base_model: codellama/CodeLlama-34b-hf model_type: LlamaForCausalLM tokenizer_type: CodeLlamaTokenizer is_llama_derived_model: true load_in_8bit: false load_in_4bit: true strict: false datasets: - path: iamtarun/code_instructions_120k_alpaca type: alpaca dataset_prepared_path: val_set_size: 0.05 output_dir: ./maverick34b adapter: qlora lora_model_dir: sequence_len: 4096 sample_packing: true pad_to_sequence_len: true lora_r: 32 lora_alpha: 16 lora_dropout: 0.05 lora_target_modules: lora_target_linear: true lora_fan_in_fan_out: wandb_project: wandb_entity: wandb_watch: wandb_name: wandb_log_model: gradient_accumulation_steps: 4 micro_batch_size: 2 num_epochs: 4 optimizer: paged_adamw_32bit lr_scheduler: cosine learning_rate: 0.0002 train_on_inputs: false group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true early_stopping_patience: resume_from_checkpoint: local_rank: logging_steps: 1 xformers_attention: flash_attention: true warmup_steps: 10 evals_per_epoch: 4 saves_per_epoch: 1 debug: deepspeed: weight_decay: 0.0 fsdp: fsdp_config: special_tokens: bos_token: "<s>" eos_token: "</s>" unk_token: "<unk>" ``` </details><br> # maverick34b This model is a fine-tuned version of [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.3391 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - distributed_type: multi-GPU - num_devices: 7 - gradient_accumulation_steps: 4 - total_train_batch_size: 56 - total_eval_batch_size: 14 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 10 - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.5065 | 0.01 | 1 | 0.5089 | | 0.3477 | 0.25 | 29 | 0.3561 | | 0.3593 | 0.51 | 58 | 0.3461 | | 0.3329 | 0.76 | 87 | 0.3423 | | 0.3607 | 1.0 | 116 | 0.3404 | | 0.3336 | 1.26 | 145 | 0.3395 | | 0.3449 | 1.51 | 174 | 0.3386 | | 0.3187 | 1.77 | 203 | 0.3377 | | 0.3216 | 2.0 | 232 | 0.3371 | | 0.2961 | 2.26 | 261 | 0.3380 | | 0.3117 | 2.51 | 290 | 0.3381 | | 0.3207 | 2.77 | 319 | 0.3379 | | 0.3047 | 3.01 | 348 | 0.3376 | | 0.3096 | 3.26 | 377 | 0.3391 | | 0.3148 | 3.52 | 406 | 0.3391 | | 0.3116 | 3.77 | 435 | 0.3391 | ### Framework versions - PEFT 0.8.2 - Transformers 4.38.0.dev0 - Pytorch 2.0.1+cu118 - Datasets 2.17.0 - Tokenizers 0.15.0
sunyijia97/falcon-7b-qlora-cstuqa-v4
sunyijia97
2024-02-17T07:28:33Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-02-17T07:28:28Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
reachrkr/falcon-rw-1bt-gptq-2bit-ptb
reachrkr
2024-02-17T07:17:04Z
6
0
transformers
[ "transformers", "safetensors", "falcon", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "2-bit", "gptq", "region:us" ]
text-generation
2024-02-17T07:16:38Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Ricosama/outputM
Ricosama
2024-02-17T06:58:13Z
1
0
peft
[ "peft", "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:mistralai/Mixtral-8x7B-Instruct-v0.1", "base_model:adapter:mistralai/Mixtral-8x7B-Instruct-v0.1", "license:apache-2.0", "region:us" ]
null
2024-02-17T06:06:29Z
--- license: apache-2.0 library_name: peft tags: - trl - sft - generated_from_trainer base_model: mistralai/Mixtral-8x7B-Instruct-v0.1 model-index: - name: outputM results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # outputM This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.00025 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 0.03 - num_epochs: 1 ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
RichardKhanhWin/dqn-SpaceInvadersNoFrameskip-v4
RichardKhanhWin
2024-02-17T06:40:19Z
0
0
stable-baselines3
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2024-02-17T06:39:46Z
--- library_name: stable-baselines3 tags: - SpaceInvadersNoFrameskip-v4 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: DQN results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: SpaceInvadersNoFrameskip-v4 type: SpaceInvadersNoFrameskip-v4 metrics: - type: mean_reward value: 540.50 +/- 74.45 name: mean_reward verified: false --- # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib Install the RL Zoo (with SB3 and SB3-Contrib): ```bash pip install rl_zoo3 ``` ``` # Download model and save it into the logs/ folder python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga RichardKhanhWin -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do: ``` python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga RichardKhanhWin -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga RichardKhanhWin ``` ## Hyperparameters ```python OrderedDict([('batch_size', 32), ('buffer_size', 100000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 4), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 100000), ('n_timesteps', 1000000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 4), ('normalize', False)]) ``` # Environment Arguments ```python {'render_mode': 'rgb_array'} ```
sunyijia97/falcon-7b-qlora-cstuqa-v3
sunyijia97
2024-02-17T06:35:03Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-02-17T06:17:28Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
gregor160300/llama2-fine-tuned-deny-sql-1-epoch
gregor160300
2024-02-17T06:25:04Z
3
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-17T06:22:05Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
VenkateshSoni/roberta-finetuned-Med
VenkateshSoni
2024-02-17T06:12:55Z
13
0
transformers
[ "transformers", "tensorboard", "safetensors", "roberta", "question-answering", "generated_from_trainer", "base_model:deepset/roberta-base-squad2", "base_model:finetune:deepset/roberta-base-squad2", "license:cc-by-4.0", "endpoints_compatible", "region:us" ]
question-answering
2024-02-17T05:15:21Z
--- license: cc-by-4.0 base_model: deepset/roberta-base-squad2 tags: - generated_from_trainer model-index: - name: roberta-finetuned-Med results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-finetuned-Med This model is a fine-tuned version of [deepset/roberta-base-squad2](https://huggingface.co/deepset/roberta-base-squad2) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
callmesan/hindi_base_wav2vec2-audio-abuse-feature
callmesan
2024-02-17T06:03:51Z
1
0
transformers
[ "transformers", "pytorch", "wav2vec2", "audio-classification", "generated_from_trainer", "base_model:Harveenchadha/hindi_base_wav2vec2", "base_model:finetune:Harveenchadha/hindi_base_wav2vec2", "license:apache-2.0", "endpoints_compatible", "region:us" ]
audio-classification
2024-02-17T05:32:41Z
--- license: apache-2.0 base_model: Harveenchadha/hindi_base_wav2vec2 tags: - generated_from_trainer metrics: - accuracy model-index: - name: hindi_base_wav2vec2-audio-abuse-feature results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # hindi_base_wav2vec2-audio-abuse-feature This model is a fine-tuned version of [Harveenchadha/hindi_base_wav2vec2](https://huggingface.co/Harveenchadha/hindi_base_wav2vec2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7202 - Accuracy: 0.6694 - Macro F1-score: 0.6693 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 50 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Macro F1-score | |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------------:| | 6.6553 | 0.77 | 10 | 6.6322 | 0.0 | 0.0 | | 6.5758 | 1.54 | 20 | 6.4417 | 0.5447 | 0.2151 | | 6.3599 | 2.31 | 30 | 6.1486 | 0.5122 | 0.3621 | | 6.0708 | 3.08 | 40 | 5.7751 | 0.5041 | 0.3351 | | 5.8361 | 3.85 | 50 | 5.4662 | 0.5041 | 0.3351 | | 5.5167 | 4.62 | 60 | 5.2127 | 0.5041 | 0.3351 | | 5.289 | 5.38 | 70 | 4.9640 | 0.5041 | 0.3351 | | 5.0266 | 6.15 | 80 | 4.7282 | 0.5041 | 0.3351 | | 4.78 | 6.92 | 90 | 4.5006 | 0.5041 | 0.3351 | | 4.6197 | 7.69 | 100 | 4.2787 | 0.5041 | 0.3351 | | 4.3798 | 8.46 | 110 | 4.0506 | 0.5041 | 0.3351 | | 4.2651 | 9.23 | 120 | 3.8315 | 0.5041 | 0.3351 | | 3.9832 | 10.0 | 130 | 3.6034 | 0.5041 | 0.3351 | | 3.7163 | 10.77 | 140 | 3.3782 | 0.5041 | 0.3351 | | 3.5481 | 11.54 | 150 | 3.1510 | 0.5041 | 0.3351 | | 3.305 | 12.31 | 160 | 2.9279 | 0.5041 | 0.3351 | | 3.1589 | 13.08 | 170 | 2.7102 | 0.5041 | 0.3351 | | 2.8368 | 13.85 | 180 | 2.4942 | 0.5041 | 0.3351 | | 2.5875 | 14.62 | 190 | 2.2896 | 0.5041 | 0.3351 | | 2.5938 | 15.38 | 200 | 2.0940 | 0.5041 | 0.3351 | | 2.2346 | 16.15 | 210 | 1.9083 | 0.5041 | 0.3351 | | 2.0404 | 16.92 | 220 | 1.7372 | 0.5041 | 0.3351 | | 1.8744 | 17.69 | 230 | 1.5755 | 0.5041 | 0.3351 | | 1.6581 | 18.46 | 240 | 1.4332 | 0.5041 | 0.3351 | | 1.7251 | 19.23 | 250 | 1.3152 | 0.5041 | 0.3351 | | 1.4569 | 20.0 | 260 | 1.2093 | 0.5041 | 0.3351 | | 1.3718 | 20.77 | 270 | 1.1160 | 0.5041 | 0.3351 | | 1.1743 | 21.54 | 280 | 1.0209 | 0.5041 | 0.3351 | | 1.0744 | 22.31 | 290 | 0.9585 | 0.6585 | 0.6309 | | 1.0933 | 23.08 | 300 | 0.8902 | 0.7019 | 0.6941 | | 0.9348 | 23.85 | 310 | 0.8504 | 0.6992 | 0.6940 | | 0.9611 | 24.62 | 320 | 0.8094 | 0.6911 | 0.6901 | | 0.8307 | 25.38 | 330 | 0.7750 | 0.6992 | 0.6992 | | 0.7863 | 26.15 | 340 | 0.7776 | 0.6802 | 0.6724 | | 0.7431 | 26.92 | 350 | 0.7624 | 0.6829 | 0.6737 | | 0.7607 | 27.69 | 360 | 0.7450 | 0.6775 | 0.6747 | | 0.8054 | 28.46 | 370 | 0.7161 | 0.6938 | 0.6914 | | 0.752 | 29.23 | 380 | 0.7021 | 0.6965 | 0.6946 | | 0.72 | 30.0 | 390 | 0.7060 | 0.6856 | 0.6846 | | 0.7252 | 30.77 | 400 | 0.6968 | 0.6911 | 0.6910 | | 0.6497 | 31.54 | 410 | 0.7016 | 0.6911 | 0.6905 | | 0.6215 | 32.31 | 420 | 0.7209 | 0.6856 | 0.6848 | | 0.6143 | 33.08 | 430 | 0.6941 | 0.6856 | 0.6856 | | 0.6778 | 33.85 | 440 | 0.6887 | 0.6856 | 0.6850 | | 0.6027 | 34.62 | 450 | 0.7010 | 0.6992 | 0.6990 | | 0.6644 | 35.38 | 460 | 0.7009 | 0.6721 | 0.6674 | | 0.6178 | 36.15 | 470 | 0.6840 | 0.7019 | 0.6985 | | 0.5817 | 36.92 | 480 | 0.6974 | 0.6829 | 0.6827 | | 0.5876 | 37.69 | 490 | 0.6914 | 0.6802 | 0.6801 | | 0.5474 | 38.46 | 500 | 0.7056 | 0.6856 | 0.6855 | | 0.5327 | 39.23 | 510 | 0.7128 | 0.6802 | 0.6800 | | 0.5648 | 40.0 | 520 | 0.7067 | 0.6748 | 0.6730 | | 0.6163 | 40.77 | 530 | 0.6804 | 0.6721 | 0.6721 | | 0.514 | 41.54 | 540 | 0.6965 | 0.6775 | 0.6774 | | 0.5817 | 42.31 | 550 | 0.7177 | 0.6775 | 0.6767 | | 0.5345 | 43.08 | 560 | 0.7136 | 0.6775 | 0.6772 | | 0.525 | 43.85 | 570 | 0.7159 | 0.6883 | 0.6876 | | 0.5043 | 44.62 | 580 | 0.7110 | 0.6802 | 0.6801 | | 0.5418 | 45.38 | 590 | 0.7149 | 0.6748 | 0.6746 | | 0.5129 | 46.15 | 600 | 0.7108 | 0.6694 | 0.6694 | | 0.5331 | 46.92 | 610 | 0.7118 | 0.6667 | 0.6667 | | 0.6061 | 47.69 | 620 | 0.7248 | 0.6802 | 0.6795 | | 0.5551 | 48.46 | 630 | 0.7196 | 0.6694 | 0.6694 | | 0.5049 | 49.23 | 640 | 0.7190 | 0.6640 | 0.6638 | | 0.4663 | 50.0 | 650 | 0.7202 | 0.6694 | 0.6693 | ### Framework versions - Transformers 4.33.0 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3
sonthenguyen/OpenHermes-2.5-Mistral-7B-mt-bench-DPO-original-v3
sonthenguyen
2024-02-17T05:54:06Z
8
2
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-16T16:15:20Z
--- library_name: transformers license: apache-2.0 --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
janhq/laser-dolphin-mixtral-2x7b-dpo-GGUF
janhq
2024-02-17T05:39:46Z
9
0
transformers
[ "transformers", "gguf", "base_model:macadeliccc/laser-dolphin-mixtral-2x7b-dpo", "base_model:quantized:macadeliccc/laser-dolphin-mixtral-2x7b-dpo", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-02-16T01:08:41Z
--- license: apache-2.0 library_name: transformers base_model: macadeliccc/laser-dolphin-mixtral-2x7b-dpo model_creator: macadeliccc model_name: laser-dolphin-mixtral-2x7b-dpo quantized_by: JanHQ --- <!-- header start --> <!-- 200823 --> <div style="width: auto; margin-left: auto; margin-right: auto"> <img src="https://github.com/janhq/jan/assets/89722390/35daac7d-b895-487c-a6ac-6663daaad78e" alt="Jan banner" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <p align="center"> <a href="https://jan.ai/">Jan</a> - <a href="https://discord.gg/AsJ8krTT3N">Discord</a> </p> <!-- header end --> # Model Description This is a GGUF version of [macadeliccc/laser-dolphin-mixtral-2x7b-dpo](https://huggingface.co/macadeliccc/laser-dolphin-mixtral-2x7b-dpo) - Model creator: [macadeliccc](https://huggingface.co/macadeliccc) - Original model: [laser-dolphin-mixtral-2x7b-dpo](https://huggingface.co/macadeliccc/laser-dolphin-mixtral-2x7b-dpo) - Model description: [Readme](https://huggingface.co/macadeliccc/laser-dolphin-mixtral-2x7b-dpo/blob/main/README.md) # About Jan Jan believes in the need for an open-source AI ecosystem and is building the infra and tooling to allow open-source AIs to compete on a level playing field with proprietary ones. Jan's long-term vision is to build a cognitive framework for future robots, who are practical, useful assistants for humans and businesses in everyday life. # Jan Model Converter This is a repository for the [open-source converter](https://github.com/janhq/model-converter. We would be grateful if the community could contribute and strengthen this repository. We are aiming to expand the repo that can convert into various types of format
FINNUMBER/Yi-Ko-6B-Finch-SA-800-per400-epoch8
FINNUMBER
2024-02-17T05:38:48Z
3
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-16T16:34:58Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Zeen0/lab1_finetuning
Zeen0
2024-02-17T05:33:41Z
4
0
transformers
[ "transformers", "safetensors", "marian", "text2text-generation", "generated_from_trainer", "dataset:kde4", "base_model:Helsinki-NLP/opus-mt-en-fr", "base_model:finetune:Helsinki-NLP/opus-mt-en-fr", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-02-17T05:33:41Z
--- license: apache-2.0 base_model: Helsinki-NLP/opus-mt-en-fr tags: - generated_from_trainer datasets: - kde4 metrics: - bleu model-index: - name: marian-finetuned-kde4-en-to-fr results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: kde4 type: kde4 config: en-fr split: train args: en-fr metrics: - name: Bleu type: bleu value: 52.88398487672078 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # marian-finetuned-kde4-en-to-fr This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset. It achieves the following results on the evaluation set: - Loss: 0.8556 - Bleu: 52.8840 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
callmesan/vakyansh-wav2vec2-gujarati-gnm-100-audio-abuse-feature
callmesan
2024-02-17T05:17:48Z
1
0
transformers
[ "transformers", "pytorch", "wav2vec2", "audio-classification", "generated_from_trainer", "base_model:Harveenchadha/vakyansh-wav2vec2-gujarati-gnm-100", "base_model:finetune:Harveenchadha/vakyansh-wav2vec2-gujarati-gnm-100", "endpoints_compatible", "region:us" ]
audio-classification
2024-02-17T04:57:48Z
--- base_model: Harveenchadha/vakyansh-wav2vec2-gujarati-gnm-100 tags: - generated_from_trainer metrics: - accuracy model-index: - name: vakyansh-wav2vec2-gujarati-gnm-100-audio-abuse-feature results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vakyansh-wav2vec2-gujarati-gnm-100-audio-abuse-feature This model is a fine-tuned version of [Harveenchadha/vakyansh-wav2vec2-gujarati-gnm-100](https://huggingface.co/Harveenchadha/vakyansh-wav2vec2-gujarati-gnm-100) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6313 - Accuracy: 0.7403 - Macro F1-score: 0.6830 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 50 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Macro F1-score | |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------------:| | 6.6694 | 0.77 | 10 | 6.6451 | 0.0387 | 0.0021 | | 6.6244 | 1.54 | 20 | 6.5275 | 0.6878 | 0.0694 | | 6.4955 | 2.31 | 30 | 6.2972 | 0.7044 | 0.4133 | | 6.2586 | 3.08 | 40 | 5.9826 | 0.7044 | 0.4133 | | 6.044 | 3.85 | 50 | 5.6760 | 0.7044 | 0.4133 | | 5.7859 | 4.62 | 60 | 5.3680 | 0.7044 | 0.4133 | | 5.506 | 5.38 | 70 | 5.0967 | 0.7044 | 0.4133 | | 5.2115 | 6.15 | 80 | 4.8565 | 0.7044 | 0.4133 | | 5.0439 | 6.92 | 90 | 4.6328 | 0.7044 | 0.4133 | | 4.924 | 7.69 | 100 | 4.4207 | 0.7044 | 0.4133 | | 4.5905 | 8.46 | 110 | 4.2046 | 0.7044 | 0.4133 | | 4.4629 | 9.23 | 120 | 3.9881 | 0.7044 | 0.4133 | | 4.2224 | 10.0 | 130 | 3.7741 | 0.7044 | 0.4133 | | 4.0429 | 10.77 | 140 | 3.5620 | 0.7044 | 0.4133 | | 3.8484 | 11.54 | 150 | 3.3434 | 0.7044 | 0.4133 | | 3.6943 | 12.31 | 160 | 3.1294 | 0.7044 | 0.4133 | | 3.4667 | 13.08 | 170 | 2.9148 | 0.7044 | 0.4133 | | 3.1164 | 13.85 | 180 | 2.7000 | 0.7044 | 0.4133 | | 2.9152 | 14.62 | 190 | 2.4912 | 0.7044 | 0.4133 | | 2.7946 | 15.38 | 200 | 2.2933 | 0.7044 | 0.4133 | | 2.5293 | 16.15 | 210 | 2.1013 | 0.7044 | 0.4133 | | 2.3488 | 16.92 | 220 | 1.9167 | 0.7044 | 0.4133 | | 2.2396 | 17.69 | 230 | 1.7418 | 0.7044 | 0.4133 | | 2.0293 | 18.46 | 240 | 1.5833 | 0.7044 | 0.4133 | | 1.8431 | 19.23 | 250 | 1.4364 | 0.7044 | 0.4133 | | 1.6658 | 20.0 | 260 | 1.3038 | 0.7044 | 0.4133 | | 1.5557 | 20.77 | 270 | 1.1904 | 0.7044 | 0.4133 | | 1.3412 | 21.54 | 280 | 1.0912 | 0.7044 | 0.4133 | | 1.2984 | 22.31 | 290 | 0.9999 | 0.7044 | 0.4133 | | 1.2517 | 23.08 | 300 | 0.9240 | 0.7044 | 0.4133 | | 1.2419 | 23.85 | 310 | 0.8693 | 0.7044 | 0.4133 | | 1.0371 | 24.62 | 320 | 0.8206 | 0.7044 | 0.4133 | | 0.922 | 25.38 | 330 | 0.7805 | 0.7044 | 0.4133 | | 0.8833 | 26.15 | 340 | 0.7281 | 0.7044 | 0.4133 | | 0.9064 | 26.92 | 350 | 0.6964 | 0.7210 | 0.4922 | | 0.7483 | 27.69 | 360 | 0.6807 | 0.7569 | 0.6771 | | 0.7677 | 28.46 | 370 | 0.6561 | 0.7762 | 0.6848 | | 0.7107 | 29.23 | 380 | 0.6450 | 0.7486 | 0.6847 | | 0.7144 | 30.0 | 390 | 0.6669 | 0.7182 | 0.6808 | | 0.6656 | 30.77 | 400 | 0.6288 | 0.7486 | 0.6764 | | 0.6896 | 31.54 | 410 | 0.6029 | 0.7652 | 0.6635 | | 0.6715 | 32.31 | 420 | 0.6152 | 0.7486 | 0.7021 | | 0.6375 | 33.08 | 430 | 0.6008 | 0.7597 | 0.6966 | | 0.6342 | 33.85 | 440 | 0.5941 | 0.7652 | 0.6892 | | 0.5992 | 34.62 | 450 | 0.6102 | 0.7459 | 0.6879 | | 0.623 | 35.38 | 460 | 0.5906 | 0.7652 | 0.6914 | | 0.5489 | 36.15 | 470 | 0.5970 | 0.7624 | 0.6610 | | 0.5553 | 36.92 | 480 | 0.6324 | 0.7320 | 0.6902 | | 0.5514 | 37.69 | 490 | 0.5974 | 0.7514 | 0.6852 | | 0.5342 | 38.46 | 500 | 0.6077 | 0.7541 | 0.6954 | | 0.5337 | 39.23 | 510 | 0.6081 | 0.7514 | 0.6872 | | 0.4809 | 40.0 | 520 | 0.6685 | 0.6961 | 0.6572 | | 0.4985 | 40.77 | 530 | 0.6262 | 0.7348 | 0.6798 | | 0.4888 | 41.54 | 540 | 0.6358 | 0.7403 | 0.6773 | | 0.4737 | 42.31 | 550 | 0.6137 | 0.7624 | 0.6911 | | 0.5249 | 43.08 | 560 | 0.6456 | 0.7293 | 0.6784 | | 0.5049 | 43.85 | 570 | 0.6503 | 0.7210 | 0.6694 | | 0.4927 | 44.62 | 580 | 0.6294 | 0.7348 | 0.6663 | | 0.4553 | 45.38 | 590 | 0.6130 | 0.7541 | 0.6835 | | 0.4631 | 46.15 | 600 | 0.6524 | 0.7238 | 0.6718 | | 0.5969 | 46.92 | 610 | 0.6233 | 0.7431 | 0.6817 | | 0.4679 | 47.69 | 620 | 0.6306 | 0.7403 | 0.6848 | | 0.4932 | 48.46 | 630 | 0.6245 | 0.7486 | 0.6922 | | 0.4723 | 49.23 | 640 | 0.6304 | 0.7431 | 0.6872 | | 0.4636 | 50.0 | 650 | 0.6313 | 0.7403 | 0.6830 | ### Framework versions - Transformers 4.33.0 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3
coloteong/lab1_finetuning
coloteong
2024-02-17T05:10:33Z
12
0
transformers
[ "transformers", "safetensors", "marian", "text2text-generation", "generated_from_trainer", "dataset:kde4", "base_model:Helsinki-NLP/opus-mt-en-fr", "base_model:finetune:Helsinki-NLP/opus-mt-en-fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-02-13T02:36:03Z
--- license: apache-2.0 base_model: Helsinki-NLP/opus-mt-en-fr tags: - generated_from_trainer datasets: - kde4 model-index: - name: marian-finetuned-kde4-en-to-fr results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # marian-finetuned-kde4-en-to-fr This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
cloudyu/Pluto_24B_lora_chat
cloudyu
2024-02-17T05:06:56Z
1
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:cloudyu/Mixtral_7Bx4_MOE_DPO", "base_model:adapter:cloudyu/Mixtral_7Bx4_MOE_DPO", "license:mit", "region:us" ]
null
2024-02-17T04:58:33Z
--- library_name: peft base_model: cloudyu/Pluto_24B_DPO_200 license: mit --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
ruru2701/filmbertv1
ruru2701
2024-02-17T04:58:25Z
5
0
transformers
[ "transformers", "safetensors", "bert", "text-classification", "generated_from_trainer", "base_model:google-bert/bert-base-cased", "base_model:finetune:google-bert/bert-base-cased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T04:57:44Z
--- license: apache-2.0 base_model: bert-base-cased tags: - generated_from_trainer model-index: - name: filmbertv1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # filmbertv1 This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1522 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
ebotwick/cats_vs_dogs_image_recog_5k
ebotwick
2024-02-17T04:56:49Z
5
0
transformers
[ "transformers", "safetensors", "vit", "image-classification", "generated_from_trainer", "dataset:cats_vs_dogs", "base_model:google/vit-base-patch16-224-in21k", "base_model:finetune:google/vit-base-patch16-224-in21k", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2024-02-17T03:31:16Z
--- license: apache-2.0 base_model: google/vit-base-patch16-224-in21k tags: - generated_from_trainer datasets: - cats_vs_dogs model-index: - name: cats_vs_dogs_image_recog_5k results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # cats_vs_dogs_image_recog_5k This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the cats_vs_dogs dataset. It achieves the following results on the evaluation set: - Loss: 0.6933 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.005 - train_batch_size: 128 - eval_batch_size: 128 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.6894 | 1.0 | 46 | 0.6933 | ### Framework versions - Transformers 4.35.0 - Pytorch 2.2.0 - Datasets 2.14.6 - Tokenizers 0.14.1
callmesan/vakyansh-wav2vec2-bengali-bnm-200-audio-abuse-feature
callmesan
2024-02-17T04:55:08Z
3
0
transformers
[ "transformers", "pytorch", "wav2vec2", "audio-classification", "generated_from_trainer", "base_model:Harveenchadha/vakyansh-wav2vec2-bengali-bnm-200", "base_model:finetune:Harveenchadha/vakyansh-wav2vec2-bengali-bnm-200", "endpoints_compatible", "region:us" ]
audio-classification
2024-02-17T04:35:12Z
--- base_model: Harveenchadha/vakyansh-wav2vec2-bengali-bnm-200 tags: - generated_from_trainer metrics: - accuracy model-index: - name: vakyansh-wav2vec2-bengali-bnm-200-audio-abuse-feature results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vakyansh-wav2vec2-bengali-bnm-200-audio-abuse-feature This model is a fine-tuned version of [Harveenchadha/vakyansh-wav2vec2-bengali-bnm-200](https://huggingface.co/Harveenchadha/vakyansh-wav2vec2-bengali-bnm-200) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.8024 - Accuracy: 0.6459 - Macro F1-score: 0.6339 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 50 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Macro F1-score | |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------------:| | 6.7346 | 0.77 | 10 | 6.7291 | 0.0 | 0.0 | | 6.6926 | 1.54 | 20 | 6.6304 | 0.0027 | 0.0004 | | 6.5587 | 2.31 | 30 | 6.4243 | 0.5730 | 0.2830 | | 6.3449 | 3.08 | 40 | 6.1051 | 0.5649 | 0.5571 | | 6.1232 | 3.85 | 50 | 5.7862 | 0.4216 | 0.3430 | | 5.8191 | 4.62 | 60 | 5.5131 | 0.4027 | 0.2908 | | 5.592 | 5.38 | 70 | 5.2719 | 0.5216 | 0.5022 | | 5.3414 | 6.15 | 80 | 5.0558 | 0.6189 | 0.6186 | | 5.1331 | 6.92 | 90 | 4.8552 | 0.6865 | 0.6852 | | 4.98 | 7.69 | 100 | 4.6603 | 0.6568 | 0.6567 | | 4.7844 | 8.46 | 110 | 4.4634 | 0.6703 | 0.6702 | | 4.7028 | 9.23 | 120 | 4.2715 | 0.6568 | 0.6567 | | 4.4476 | 10.0 | 130 | 4.0733 | 0.6297 | 0.6280 | | 4.2098 | 10.77 | 140 | 3.8749 | 0.6108 | 0.6041 | | 4.0715 | 11.54 | 150 | 3.6803 | 0.5027 | 0.4564 | | 3.8545 | 12.31 | 160 | 3.4603 | 0.6649 | 0.6648 | | 3.708 | 13.08 | 170 | 3.2559 | 0.6541 | 0.6534 | | 3.4318 | 13.85 | 180 | 3.0493 | 0.6676 | 0.6675 | | 3.1874 | 14.62 | 190 | 2.8456 | 0.6838 | 0.6837 | | 3.1887 | 15.38 | 200 | 2.6625 | 0.5595 | 0.5384 | | 2.8359 | 16.15 | 210 | 2.4679 | 0.5757 | 0.5604 | | 2.6265 | 16.92 | 220 | 2.2662 | 0.6892 | 0.6841 | | 2.4536 | 17.69 | 230 | 2.0843 | 0.6649 | 0.6644 | | 2.2288 | 18.46 | 240 | 1.9218 | 0.6459 | 0.6431 | | 2.2955 | 19.23 | 250 | 1.7633 | 0.6595 | 0.6578 | | 1.9739 | 20.0 | 260 | 1.6105 | 0.6730 | 0.6671 | | 1.8575 | 20.77 | 270 | 1.4855 | 0.6378 | 0.6351 | | 1.607 | 21.54 | 280 | 1.3582 | 0.6649 | 0.6646 | | 1.4831 | 22.31 | 290 | 1.2425 | 0.6676 | 0.6646 | | 1.4484 | 23.08 | 300 | 1.1522 | 0.6703 | 0.6660 | | 1.2517 | 23.85 | 310 | 1.0688 | 0.6595 | 0.6554 | | 1.2793 | 24.62 | 320 | 1.0006 | 0.6541 | 0.6523 | | 1.0722 | 25.38 | 330 | 0.9486 | 0.6568 | 0.6543 | | 0.9888 | 26.15 | 340 | 0.9292 | 0.6135 | 0.6135 | | 0.9134 | 26.92 | 350 | 0.8580 | 0.6514 | 0.6492 | | 0.9208 | 27.69 | 360 | 0.8352 | 0.6649 | 0.6646 | | 0.966 | 28.46 | 370 | 0.8220 | 0.6162 | 0.6160 | | 0.8746 | 29.23 | 380 | 0.8064 | 0.6568 | 0.6420 | | 0.8619 | 30.0 | 390 | 0.7856 | 0.6405 | 0.5942 | | 0.841 | 30.77 | 400 | 0.7612 | 0.6459 | 0.6020 | | 0.7629 | 31.54 | 410 | 0.7441 | 0.6459 | 0.6434 | | 0.6736 | 32.31 | 420 | 0.7610 | 0.6568 | 0.6562 | | 0.6579 | 33.08 | 430 | 0.7624 | 0.6514 | 0.6456 | | 0.7514 | 33.85 | 440 | 0.7374 | 0.6649 | 0.6467 | | 0.6579 | 34.62 | 450 | 0.7503 | 0.6541 | 0.6471 | | 0.6864 | 35.38 | 460 | 0.8286 | 0.5892 | 0.5889 | | 0.6863 | 36.15 | 470 | 0.7393 | 0.6541 | 0.6396 | | 0.6224 | 36.92 | 480 | 0.7427 | 0.6541 | 0.6507 | | 0.6255 | 37.69 | 490 | 0.7495 | 0.6405 | 0.6268 | | 0.5295 | 38.46 | 500 | 0.7787 | 0.6486 | 0.6385 | | 0.5549 | 39.23 | 510 | 0.7909 | 0.6378 | 0.6360 | | 0.5752 | 40.0 | 520 | 0.7631 | 0.6459 | 0.6361 | | 0.616 | 40.77 | 530 | 0.7636 | 0.6432 | 0.6390 | | 0.5038 | 41.54 | 540 | 0.7847 | 0.6514 | 0.6372 | | 0.5935 | 42.31 | 550 | 0.7837 | 0.6595 | 0.6461 | | 0.5453 | 43.08 | 560 | 0.7804 | 0.6405 | 0.6330 | | 0.5378 | 43.85 | 570 | 0.7928 | 0.6514 | 0.6338 | | 0.4852 | 44.62 | 580 | 0.8249 | 0.6324 | 0.6285 | | 0.5198 | 45.38 | 590 | 0.8065 | 0.6459 | 0.6186 | | 0.5067 | 46.15 | 600 | 0.8210 | 0.6162 | 0.6107 | | 0.5533 | 46.92 | 610 | 0.8053 | 0.6432 | 0.6300 | | 0.6282 | 47.69 | 620 | 0.7970 | 0.6459 | 0.6316 | | 0.5617 | 48.46 | 630 | 0.8095 | 0.6243 | 0.6165 | | 0.5016 | 49.23 | 640 | 0.8038 | 0.6378 | 0.6274 | | 0.467 | 50.0 | 650 | 0.8024 | 0.6459 | 0.6339 | ### Framework versions - Transformers 4.33.0 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3
sunyijia97/falcon-7b-qlora-chat-support-bot-faq
sunyijia97
2024-02-17T04:54:11Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-02-17T04:54:08Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
fzzhang/toten_gsm8k_s
fzzhang
2024-02-17T04:52:45Z
5
0
peft
[ "peft", "tensorboard", "safetensors", "generated_from_trainer", "base_model:Toten5/Marcoroni-neural-chat-7B-v2", "base_model:adapter:Toten5/Marcoroni-neural-chat-7B-v2", "license:apache-2.0", "region:us" ]
null
2024-02-16T22:20:48Z
--- license: apache-2.0 library_name: peft tags: - generated_from_trainer base_model: Toten5/Marcoroni-neural-chat-7B-v2 model-index: - name: toten_gsm8k_s results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # toten_gsm8k_s This model is a fine-tuned version of [Toten5/Marcoroni-neural-chat-7B-v2](https://huggingface.co/Toten5/Marcoroni-neural-chat-7B-v2) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 0 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results ### Framework versions - PEFT 0.7.2.dev0 - Transformers 4.37.0.dev0 - Pytorch 2.1.2+cu121 - Datasets 2.16.1 - Tokenizers 0.15.0
minjiyoo/linguistic-complexity-llama-2-7b-4000
minjiyoo
2024-02-17T04:37:08Z
0
0
peft
[ "peft", "region:us" ]
null
2024-01-28T02:59:10Z
--- library_name: peft --- ## Training procedure ### Framework versions - PEFT 0.4.0
Xenon1/Voyage
Xenon1
2024-02-17T04:33:46Z
3
1
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "Voyage", "conversational", "en", "arxiv:2401.10020", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-17T04:29:39Z
--- language: - en license: apache-2.0 tags: - mistral - Voyage pipeline_tag: text-generation --- # Model Card for Voyage Mistral-7B-v0.1 model fine-tuned on the Ultrafeedback dataset using techinques shown in the paper [Self-Rewarding Language Models](https://arxiv.org/abs/2401.10020). ## Results | model_name | Average | arc_challenge | hellaswag | mmlu | truthfulqa_mc2 | winogrande | |:-------------|----------:|----------------:|------------:|---------:|-----------------:|-------------:| | Voyage | 0.68526 | 0.613481 | 0.848337 | 0.595998 | 0.602897 | 0.765588 | ## Instruction format In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id. E.g. ``` text = "<s>[INST] What is your favourite condiment? [/INST]" "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> " "[INST] Do you have mayonnaise recipes? [/INST]" ``` This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method: ```python from transformers import AutoModelForCausalLM, AutoTokenizer device = "cuda" # the device to load the model onto model = AutoModelForCausalLM.from_pretrained("Xenon1/Voyage") tokenizer = AutoTokenizer.from_pretrained("Xenon1/Voyage") messages = [ {"role": "user", "content": "What is your favourite condiment?"}, {"role": "assistant", "content": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"}, {"role": "user", "content": "Do you have mayonnaise recipes?"} ] encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt") model_inputs = encodeds.to(device) model.to(device) generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True) decoded = tokenizer.batch_decode(generated_ids) print(decoded[0]) ``` ## Model Architecture This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices: - Grouped-Query Attention - Sliding-Window Attention - Byte-fallback BPE tokenizer
theidoldaily/yoshiko-tsushima
theidoldaily
2024-02-17T04:28:29Z
5
0
diffusers
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:cagliostrolab/animagine-xl-3.0", "base_model:adapter:cagliostrolab/animagine-xl-3.0", "license:mit", "region:us" ]
text-to-image
2024-02-17T04:24:27Z
--- tags: - text-to-image - stable-diffusion - lora - diffusers - template:sd-lora widget: - text: >- masterpiece, high quality, defined pupil, looking at viewer, rounded pupil, defined iris, (soft iris:1.2), parameters: negative_prompt: >- bad_anatomy, deformation, amputation, deformity, deformed_nipples, duplicated_torso, deformed_torso, long_torso, large_torso, unproportioned_torso, (deformed_pussy:1.2), (deformed_hands:1.2), unproportioned_eyes, unproportioned_head, small_head, duplicated_nose, big_nose, fusioned_clothes, fusioned_arms, undefined_limbs, divided_pussy, red_pussy, duplicated_pussy, deformed_anus, deformed_pussy, output: url: images/yoshiko_final.png base_model: cagliostrolab/animagine-xl-3.0 instance_prompt: id_yoshiko_tsushima license: mit --- # Yoshiko Tsushima <Gallery /> ## Model description This model was trained to generate high quality images based on SIFAS cards. To achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement. ## Trigger words You should use `id_yoshiko_tsushima` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/theidoldaily/yoshiko-tsushima/tree/main) them in the Files & versions tab.
kaitchup/Qwen1.5-7B-bnb-4bit
kaitchup
2024-02-17T04:24:55Z
259
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
2024-02-17T04:22:25Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
theidoldaily/hanamaru-kunikida
theidoldaily
2024-02-17T04:23:46Z
3
0
diffusers
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:cagliostrolab/animagine-xl-3.0", "base_model:adapter:cagliostrolab/animagine-xl-3.0", "license:mit", "region:us" ]
text-to-image
2024-02-17T04:21:00Z
--- tags: - text-to-image - stable-diffusion - lora - diffusers - template:sd-lora widget: - text: >- masterpiece, high quality, defined pupil, looking at viewer, rounded pupil, defined iris, (soft iris:1.2), parameters: negative_prompt: >- bad_anatomy, deformation, amputation, deformity, deformed_nipples, duplicated_torso, deformed_torso, long_torso, large_torso, unproportioned_torso, (deformed_pussy:1.2), (deformed_hands:1.2), unproportioned_eyes, unproportioned_head, small_head, duplicated_nose, big_nose, fusioned_clothes, fusioned_arms, undefined_limbs, divided_pussy, red_pussy, duplicated_pussy, deformed_anus, deformed_pussy, output: url: images/hanamaru_final.png base_model: cagliostrolab/animagine-xl-3.0 instance_prompt: id_hanamaru_kunikida license: mit --- # Hanamaru Kunikida <Gallery /> ## Model description This model was trained to generate high quality images based on SIFAS cards. To achieve better quality, you should be using hako-mikan&#39;s regional prompter, along with Latent Mode, which modifies the way Stable Diffusion isolates the LoRA resulting in a significant improvement. ## Trigger words You should use `id_hanamaru_kunikida` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/theidoldaily/hanamaru-kunikida/tree/main) them in the Files & versions tab.
Xenon1/Oasis
Xenon1
2024-02-17T04:23:27Z
5
0
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "mistral", "Oasis", "en", "arxiv:2401.10020", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-17T04:16:28Z
--- language: - en license: apache-2.0 tags: - mistral - Oasis pipeline_tag: text-generation --- # Model Card for Oasis Mistral-7B-v0.1 model fine-tuned on the Ultrafeedback dataset using techinques shown in the paper [Self-Rewarding Language Models](https://arxiv.org/abs/2401.10020). ## Results | model_name | Average | arc_challenge | gsm8k | hellaswag | mmlu | truthfulqa_mc2 | winogrande | |:-------------|----------:|----------------:|---------:|------------:|---------:|-----------------:|-------------:| | Oasis | 0.701904 | 0.613481 | 0.741471 | 0.848337 | 0.639652 | 0.602897 | 0.765588 | ## Instruction format In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id. E.g. ``` text = "<s>[INST] What is your favourite condiment? [/INST]" "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> " "[INST] Do you have mayonnaise recipes? [/INST]" ``` This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method: ```python from transformers import AutoModelForCausalLM, AutoTokenizer device = "cuda" # the device to load the model onto model = AutoModelForCausalLM.from_pretrained("Xenon1/Oasis") tokenizer = AutoTokenizer.from_pretrained("Xenon1/Oasis") messages = [ {"role": "user", "content": "What is your favourite condiment?"}, {"role": "assistant", "content": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"}, {"role": "user", "content": "Do you have mayonnaise recipes?"} ] encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt") model_inputs = encodeds.to(device) model.to(device) generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True) decoded = tokenizer.batch_decode(generated_ids) print(decoded[0]) ``` ## Model Architecture This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices: - Grouped-Query Attention - Sliding-Window Attention - Byte-fallback BPE tokenizer
VladS159/Whisper_medium_ro_VladS_02_16_24_1500_steps_multi_gpu
VladS159
2024-02-17T04:21:05Z
3
0
transformers
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "hf-asr-leaderboard", "generated_from_trainer", "ro", "dataset:mozilla-foundation/common_voice_16_1", "base_model:openai/whisper-medium", "base_model:finetune:openai/whisper-medium", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2024-02-16T12:31:05Z
--- language: - ro license: apache-2.0 base_model: openai/whisper-medium tags: - hf-asr-leaderboard - generated_from_trainer datasets: - mozilla-foundation/common_voice_16_1 metrics: - wer model-index: - name: Whisper Medium Ro - Sarbu Vlad results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: Common Voice 16.1 type: mozilla-foundation/common_voice_16_1 args: 'config: ro, split: test' metrics: - name: Wer type: wer value: 12.10333666091902 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Whisper Medium Ro - Sarbu Vlad This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the Common Voice 16.1 dataset. It achieves the following results on the evaluation set: - Loss: 0.1719 - Wer: 12.1033 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - distributed_type: multi-GPU - num_devices: 3 - total_train_batch_size: 48 - total_eval_batch_size: 48 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 250 - training_steps: 2500 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.1563 | 0.98 | 250 | 0.1542 | 14.6716 | | 0.0933 | 1.96 | 500 | 0.1306 | 13.0714 | | 0.0428 | 2.94 | 750 | 0.1298 | 11.8886 | | 0.0243 | 3.92 | 1000 | 0.1353 | 12.0096 | | 0.0147 | 4.9 | 1250 | 0.1433 | 12.1064 | | 0.0083 | 5.88 | 1500 | 0.1572 | 12.2606 | | 0.0052 | 6.86 | 1750 | 0.1591 | 12.3090 | | 0.0037 | 7.84 | 2000 | 0.1665 | 12.0307 | | 0.0026 | 8.82 | 2250 | 0.1708 | 12.0549 | | 0.0021 | 9.8 | 2500 | 0.1719 | 12.1033 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0 - Datasets 2.17.0 - Tokenizers 0.15.1
DouglasChan/lab1_finetuned
DouglasChan
2024-02-17T04:05:05Z
118
0
transformers
[ "transformers", "safetensors", "marian", "text2text-generation", "generated_from_trainer", "dataset:kde4", "base_model:Helsinki-NLP/opus-mt-en-fr", "base_model:finetune:Helsinki-NLP/opus-mt-en-fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-02-13T19:06:20Z
--- license: apache-2.0 base_model: Helsinki-NLP/opus-mt-en-fr tags: - generated_from_trainer datasets: - kde4 model-index: - name: marian-finetuned-kde4-en-to-fr results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # marian-finetuned-kde4-en-to-fr This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
MAdAiLab/llama2_7b_bf16_adapter_merged_final
MAdAiLab
2024-02-17T04:02:34Z
4
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-17T03:57:25Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Ernoche/N
Ernoche
2024-02-17T03:58:37Z
0
0
null
[ "license:other", "region:us" ]
null
2024-02-17T03:58:37Z
--- license: other license_name: other license_link: LICENSE ---
kaitchup/Qwen1.5-7B-gptq-4bit
kaitchup
2024-02-17T03:52:43Z
63
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "gptq", "region:us" ]
text-generation
2024-02-17T03:50:18Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
cbdb/OfficeTitleAddressSplitter
cbdb
2024-02-17T03:32:08Z
94
1
transformers
[ "transformers", "safetensors", "bert", "token-classification", "Seq2SeqLM", "古文", "文言文", "中国古代官职地名拆分", "ancient", "classical", "zh", "license:cc-by-nc-sa-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2024-01-26T22:44:13Z
--- language: - zh tags: - Seq2SeqLM - 古文 - 文言文 - 中国古代官职地名拆分 - ancient - classical license: cc-by-nc-sa-4.0 --- # <font color="IndianRed"> OTAS (Office Title Address Splitter)</font> [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1UoG3QebyBlK6diiYckiQv-5dRB9dA4iv?usp=sharing) Our model <font color="cornflowerblue">OTAS (Office Title Address Splitter) </font> is a Named Entity Recognition Classical Chinese language model that is intended to <font color="IndianRed">split the address portion in Classical Chinese office titles.</font>. This model is first inherited from raynardj/classical-chinese-punctuation-guwen-biaodian Classical Chinese punctuation model, and finetuned using over a 25,000 high-quality punctuation pairs collected CBDB group (China Biographical Database). ### <font color="IndianRed"> Sample input txt file </font> The sample input txt file can be downloaded here: https://huggingface.co/cbdb/OfficeTitleAddressSplitter/blob/main/input.txt ### <font color="IndianRed"> How to use </font> Here is how to use this model to get the features of a given text in PyTorch: <font color="cornflowerblue"> 1. Import model and packages </font> ```python from transformers import AutoTokenizer, AutoModelForTokenClassification PRETRAINED = "cbdb/OfficeTitleAddressSplitter" tokenizer = AutoTokenizer.from_pretrained(PRETRAINED) model = AutoModelForTokenClassification.from_pretrained(PRETRAINED) ``` <font color="cornflowerblue"> 2. Load Data </font> ```python # Load your data here test_list = ['漢軍鑲黃旗副都統', '兵部右侍郎', '盛京戶部侍郎'] ``` <font color="cornflowerblue"> 3. Make a prediction </font> ```python def predict_class(test): tokens_test = tokenizer.encode_plus( test, add_special_tokens=True, return_attention_mask=True, padding=True, max_length=128, return_tensors='pt', truncation=True ) test_seq = torch.tensor(tokens_test['input_ids']) test_mask = torch.tensor(tokens_test['attention_mask']) inputs = { "input_ids": test_seq, "attention_mask": test_mask } with torch.no_grad(): # print(inputs.shape) outputs = model(**inputs) outputs = outputs.logits.detach().cpu().numpy() softmax_score = softmax(outputs) softmax_score = np.argmax(softmax_score, axis=2)[0] return test_seq, softmax_score for test_sen0 in test_list: test_seq, pred_class_proba = predict_class(test_sen0) test_sen = tokenizer.decode(test_seq[0]).split() label = [idx2label[i] for i in pred_class_proba] element_to_find = '。' if element_to_find in label: index = label.index(element_to_find) test_sen_pred = [i for i in test_sen0] test_sen_pred.insert(index, element_to_find) test_sen_pred = ''.join(test_sen_pred) else: test_sen_pred = [i for i in test_sen0] test_sen_pred = ''.join(test_sen_pred) print(test_sen_pred) ``` 漢軍鑲黃旗。副都統<br> 兵部右侍郎<br> 盛京。戶部侍郎<br> ### <font color="IndianRed">Authors </font> Queenie Luo (queenieluo[at]g.harvard.edu) <br> Hongsu Wang <br> Peter Bol <br> CBDB Group ### <font color="IndianRed">License </font> Copyright (c) 2023 CBDB Except where otherwise noted, content on this repository is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0). To view a copy of this license, visit http://creativecommons.org/licenses/by-nc-sa/4.0/ or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042, USA.
B2111797/trans-en-vi-v1
B2111797
2024-02-17T03:26:41Z
120
0
transformers
[ "transformers", "safetensors", "marian", "text2text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-02-17T03:26:24Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
LoneStriker/LWM-Text-Chat-256K-GPTQ
LoneStriker
2024-02-17T03:13:52Z
4
1
transformers
[ "transformers", "pytorch", "llama", "text-generation", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
2024-02-17T03:10:40Z
--- inference: false --- <br> <br> # LWM-Text-Chat-256K Model Card ## Model details **Model type:** LWM-Text-Chat-256K is an open-source model trained from LLaMA-2 on a subset of Books3 filtered data. It is an auto-regressive language model, based on the transformer architecture. **Model date:** LWM-Text-Chat-256K was trained in December 2023. **Paper or resources for more information:** https://largeworldmodel.github.io/ ## License Llama 2 is licensed under the LLAMA 2 Community License, Copyright (c) Meta Platforms, Inc. All Rights Reserved. **Where to send questions or comments about the model:** https://github.com/LargeWorldModel/lwm/issues ## Training dataset - 37K subset of Books3 documents with 200K to 500K tokens
Shijia/furina_hau_loss_2e-05
Shijia
2024-02-17T03:13:05Z
101
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T03:12:08Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_hau_loss_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_hau_loss_2e-05 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0213 - Spearman Corr: 0.7683 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.95 | 200 | 0.0215 | 0.7652 | | No log | 1.91 | 400 | 0.0209 | 0.7689 | | 0.0007 | 2.86 | 600 | 0.0213 | 0.7713 | | 0.0007 | 3.82 | 800 | 0.0208 | 0.7714 | | 0.0006 | 4.77 | 1000 | 0.0205 | 0.7709 | | 0.0006 | 5.73 | 1200 | 0.0205 | 0.7725 | | 0.0006 | 6.68 | 1400 | 0.0212 | 0.7675 | | 0.0006 | 7.64 | 1600 | 0.0213 | 0.7675 | | 0.0005 | 8.59 | 1800 | 0.0210 | 0.7672 | | 0.0005 | 9.55 | 2000 | 0.0211 | 0.7664 | | 0.0005 | 10.5 | 2200 | 0.0216 | 0.7675 | | 0.0005 | 11.46 | 2400 | 0.0212 | 0.7666 | | 0.0005 | 12.41 | 2600 | 0.0213 | 0.7683 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
theofilusarifin/image_classification
theofilusarifin
2024-02-17T03:12:48Z
186
0
transformers
[ "transformers", "tensorboard", "safetensors", "vit", "image-classification", "generated_from_trainer", "dataset:imagefolder", "base_model:google/vit-base-patch16-224-in21k", "base_model:finetune:google/vit-base-patch16-224-in21k", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2024-02-16T16:42:30Z
--- license: apache-2.0 base_model: google/vit-base-patch16-224-in21k tags: - generated_from_trainer datasets: - imagefolder model-index: - name: image_classification results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # image_classification This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset. It achieves the following results on the evaluation set: - eval_loss: 1.3813 - eval_accuracy: 0.5312 - eval_runtime: 179.1366 - eval_samples_per_second: 0.893 - eval_steps_per_second: 0.056 - epoch: 9.43 - step: 377 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 40 ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
zhudanhao/rlcourse_u2_taxi
zhudanhao
2024-02-17T03:08:13Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2024-02-17T03:08:09Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: rlcourse_u2_taxi results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.54 +/- 2.73 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="zhudanhao/rlcourse_u2_taxi", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
Shijia/furina_pan_loss_5e-06
Shijia
2024-02-17T03:06:47Z
89
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T03:05:41Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_pan_loss_5e-06 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_pan_loss_5e-06 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0215 - Spearman Corr: 0.7688 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.85 | 200 | 0.0213 | 0.7680 | | No log | 1.69 | 400 | 0.0213 | 0.7726 | | 0.0011 | 2.54 | 600 | 0.0213 | 0.7669 | | 0.0011 | 3.38 | 800 | 0.0221 | 0.7654 | | 0.0008 | 4.23 | 1000 | 0.0221 | 0.7698 | | 0.0008 | 5.07 | 1200 | 0.0227 | 0.7682 | | 0.0008 | 5.92 | 1400 | 0.0211 | 0.7688 | | 0.0009 | 6.77 | 1600 | 0.0210 | 0.7699 | | 0.0009 | 7.61 | 1800 | 0.0212 | 0.7687 | | 0.002 | 8.46 | 2000 | 0.0213 | 0.7700 | | 0.002 | 9.3 | 2200 | 0.0216 | 0.7675 | | 0.0019 | 10.15 | 2400 | 0.0215 | 0.7692 | | 0.0019 | 10.99 | 2600 | 0.0215 | 0.7695 | | 0.0019 | 11.84 | 2800 | 0.0212 | 0.7699 | | 0.0019 | 12.68 | 3000 | 0.0214 | 0.7691 | | 0.0019 | 13.53 | 3200 | 0.0215 | 0.7688 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
zhudanhao/RlCourse
zhudanhao
2024-02-17T02:56:02Z
1
0
transformers
[ "transformers", "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "endpoints_compatible", "region:us" ]
reinforcement-learning
2024-02-16T07:49:12Z
--- tags: - FrozenLake-v1-4x4-no_slippery - q-learning - reinforcement-learning - custom-implementation model-index: - name: RlCourse results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: FrozenLake-v1-4x4-no_slippery type: FrozenLake-v1-4x4-no_slippery metrics: - type: mean_reward value: 1.00 +/- 0.00 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
Shijia/furina_hau_corr_2e-05
Shijia
2024-02-17T02:54:44Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:53:41Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_hau_corr_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_hau_corr_2e-05 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0209 - Spearman Corr: 0.7736 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.95 | 200 | 0.0214 | 0.7723 | | No log | 1.91 | 400 | 0.0225 | 0.7716 | | 0.0012 | 2.86 | 600 | 0.0211 | 0.7695 | | 0.0012 | 3.82 | 800 | 0.0207 | 0.7718 | | 0.0011 | 4.77 | 1000 | 0.0214 | 0.7723 | | 0.0011 | 5.73 | 1200 | 0.0209 | 0.7753 | | 0.001 | 6.68 | 1400 | 0.0210 | 0.7710 | | 0.001 | 7.64 | 1600 | 0.0204 | 0.7721 | | 0.0009 | 8.59 | 1800 | 0.0217 | 0.7731 | | 0.0009 | 9.55 | 2000 | 0.0216 | 0.7692 | | 0.0009 | 10.5 | 2200 | 0.0206 | 0.7724 | | 0.0009 | 11.46 | 2400 | 0.0213 | 0.7734 | | 0.0009 | 12.41 | 2600 | 0.0208 | 0.7725 | | 0.0009 | 13.37 | 2800 | 0.0207 | 0.7760 | | 0.0008 | 14.32 | 3000 | 0.0209 | 0.7724 | | 0.0008 | 15.27 | 3200 | 0.0208 | 0.7729 | | 0.0007 | 16.23 | 3400 | 0.0212 | 0.7732 | | 0.0007 | 17.18 | 3600 | 0.0209 | 0.7746 | | 0.0007 | 18.14 | 3800 | 0.0209 | 0.7745 | | 0.0007 | 19.09 | 4000 | 0.0202 | 0.7759 | | 0.0007 | 20.05 | 4200 | 0.0206 | 0.7750 | | 0.0007 | 21.0 | 4400 | 0.0209 | 0.7736 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
LoneStriker/LWM-Text-Chat-128K-AWQ
LoneStriker
2024-02-17T02:54:18Z
62
0
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "autotrain_compatible", "text-generation-inference", "4-bit", "awq", "region:us" ]
text-generation
2024-02-17T02:52:37Z
--- inference: false --- <br> <br> # LWM-Text-Chat-128K Model Card ## Model details **Model type:** LWM-Text-Chat-128K is an open-source model trained from LLaMA-2 on a subset of Books3 filtered data. It is an auto-regressive language model, based on the transformer architecture. **Model date:** LWM-Text-Chat-128K was trained in December 2023. **Paper or resources for more information:** https://largeworldmodel.github.io/ ## License Llama 2 is licensed under the LLAMA 2 Community License, Copyright (c) Meta Platforms, Inc. All Rights Reserved. **Where to send questions or comments about the model:** https://github.com/LargeWorldModel/lwm/issues ## Training dataset - 92K subset of Books3 documents with 100K to 200K tokens
andyleetw/NeuralPipe-7B-slerp
andyleetw
2024-02-17T02:51:23Z
4
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "OpenPipe/mistral-ft-optimized-1218", "mlabonne/NeuralHermes-2.5-Mistral-7B", "base_model:OpenPipe/mistral-ft-optimized-1218", "base_model:merge:OpenPipe/mistral-ft-optimized-1218", "base_model:mlabonne/NeuralHermes-2.5-Mistral-7B", "base_model:merge:mlabonne/NeuralHermes-2.5-Mistral-7B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-17T02:46:52Z
--- tags: - merge - mergekit - lazymergekit - OpenPipe/mistral-ft-optimized-1218 - mlabonne/NeuralHermes-2.5-Mistral-7B base_model: - OpenPipe/mistral-ft-optimized-1218 - mlabonne/NeuralHermes-2.5-Mistral-7B --- # NeuralPipe-7B-slerp NeuralPipe-7B-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [OpenPipe/mistral-ft-optimized-1218](https://huggingface.co/OpenPipe/mistral-ft-optimized-1218) * [mlabonne/NeuralHermes-2.5-Mistral-7B](https://huggingface.co/mlabonne/NeuralHermes-2.5-Mistral-7B) ## 🧩 Configuration ```yaml slices: - sources: - model: OpenPipe/mistral-ft-optimized-1218 layer_range: [0, 32] - model: mlabonne/NeuralHermes-2.5-Mistral-7B layer_range: [0, 32] merge_method: slerp base_model: OpenPipe/mistral-ft-optimized-1218 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "andyleetw/NeuralPipe-7B-slerp" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
Shijia/furina_kin_loss_5e-06
Shijia
2024-02-17T02:41:01Z
92
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:39:55Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_kin_loss_5e-06 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_kin_loss_5e-06 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0215 - Spearman Corr: 0.7692 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.89 | 200 | 0.0220 | 0.7641 | | No log | 1.78 | 400 | 0.0209 | 0.7696 | | 0.0013 | 2.67 | 600 | 0.0213 | 0.7674 | | 0.0013 | 3.56 | 800 | 0.0216 | 0.7691 | | 0.0013 | 4.45 | 1000 | 0.0217 | 0.7698 | | 0.0013 | 5.35 | 1200 | 0.0218 | 0.7681 | | 0.002 | 6.24 | 1400 | 0.0219 | 0.7647 | | 0.002 | 7.13 | 1600 | 0.0218 | 0.7663 | | 0.002 | 8.02 | 1800 | 0.0210 | 0.7675 | | 0.002 | 8.91 | 2000 | 0.0215 | 0.7692 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
pravsels/deepseek-coder-6.7b-instruct-finetuned-manimation
pravsels
2024-02-17T02:39:29Z
6
0
transformers
[ "transformers", "tensorboard", "safetensors", "llama", "text-generation", "generated_from_trainer", "base_model:deepseek-ai/deepseek-coder-6.7b-instruct", "base_model:finetune:deepseek-ai/deepseek-coder-6.7b-instruct", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-01-20T21:43:19Z
--- license: other base_model: deepseek-ai/deepseek-coder-6.7b-instruct tags: - generated_from_trainer model-index: - name: deepseek-coder-6.7b-instruct-finetuned-manimation results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # deepseek-coder-6.7b-instruct-finetuned-manimation This model is a fine-tuned version of [deepseek-ai/deepseek-coder-6.7b-instruct](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instruct) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.1297 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - gradient_accumulation_steps: 16 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 0.98 | 35 | 1.1468 | | No log | 1.99 | 71 | 1.1349 | | No log | 2.95 | 105 | 1.1297 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/xlm-roberta-base_pan_loss_5e-06
Shijia
2024-02-17T02:34:58Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:34:10Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_pan_loss_5e-06 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_pan_loss_5e-06 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0245 - Spearman Corr: 0.7682 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.85 | 200 | 0.0232 | 0.7677 | | No log | 1.69 | 400 | 0.0279 | 0.7686 | | 0.0011 | 2.54 | 600 | 0.0229 | 0.7636 | | 0.0011 | 3.38 | 800 | 0.0240 | 0.7676 | | 0.0009 | 4.23 | 1000 | 0.0237 | 0.7664 | | 0.0009 | 5.07 | 1200 | 0.0239 | 0.7677 | | 0.0009 | 5.92 | 1400 | 0.0236 | 0.7678 | | 0.0007 | 6.77 | 1600 | 0.0248 | 0.7646 | | 0.0007 | 7.61 | 1800 | 0.0231 | 0.7647 | | 0.0006 | 8.46 | 2000 | 0.0256 | 0.7671 | | 0.0006 | 9.3 | 2200 | 0.0245 | 0.7682 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/xlm-roberta-base_pan_loss_2e-05
Shijia
2024-02-17T02:31:47Z
91
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:31:05Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_pan_loss_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_pan_loss_2e-05 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0216 - Spearman Corr: 0.7776 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.85 | 200 | 0.0210 | 0.7733 | | No log | 1.69 | 400 | 0.0215 | 0.7798 | | 0.0009 | 2.54 | 600 | 0.0219 | 0.7770 | | 0.0009 | 3.38 | 800 | 0.0212 | 0.7807 | | 0.0006 | 4.23 | 1000 | 0.0224 | 0.7806 | | 0.0006 | 5.07 | 1200 | 0.0210 | 0.7800 | | 0.0006 | 5.92 | 1400 | 0.0208 | 0.7799 | | 0.0004 | 6.77 | 1600 | 0.0214 | 0.7793 | | 0.0004 | 7.61 | 1800 | 0.0216 | 0.7795 | | 0.0003 | 8.46 | 2000 | 0.0207 | 0.7819 | | 0.0003 | 9.3 | 2200 | 0.0209 | 0.7826 | | 0.0004 | 10.15 | 2400 | 0.0209 | 0.7793 | | 0.0004 | 10.99 | 2600 | 0.0207 | 0.7804 | | 0.0004 | 11.84 | 2800 | 0.0210 | 0.7808 | | 0.0004 | 12.68 | 3000 | 0.0216 | 0.7776 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/furina_kin_corr_5e-06
Shijia
2024-02-17T02:31:41Z
100
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:30:39Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_kin_corr_5e-06 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_kin_corr_5e-06 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0215 - Spearman Corr: 0.7702 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.89 | 200 | 0.0217 | 0.7655 | | No log | 1.78 | 400 | 0.0210 | 0.7692 | | 0.0023 | 2.67 | 600 | 0.0215 | 0.7681 | | 0.0023 | 3.56 | 800 | 0.0214 | 0.7704 | | 0.0022 | 4.45 | 1000 | 0.0216 | 0.7710 | | 0.0022 | 5.35 | 1200 | 0.0215 | 0.7693 | | 0.0021 | 6.24 | 1400 | 0.0217 | 0.7670 | | 0.0021 | 7.13 | 1600 | 0.0217 | 0.7665 | | 0.002 | 8.02 | 1800 | 0.0209 | 0.7673 | | 0.002 | 8.91 | 2000 | 0.0215 | 0.7694 | | 0.002 | 9.8 | 2200 | 0.0212 | 0.7709 | | 0.002 | 10.69 | 2400 | 0.0218 | 0.7674 | | 0.002 | 11.58 | 2600 | 0.0215 | 0.7702 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/furina_esp_loss_2e-05
Shijia
2024-02-17T02:27:42Z
91
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:26:37Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_esp_loss_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_esp_loss_2e-05 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0213 - Spearman Corr: 0.7685 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.94 | 200 | 0.0214 | 0.7716 | | No log | 1.89 | 400 | 0.0230 | 0.7712 | | 0.0009 | 2.83 | 600 | 0.0204 | 0.7729 | | 0.0009 | 3.77 | 800 | 0.0224 | 0.7691 | | 0.001 | 4.72 | 1000 | 0.0215 | 0.7706 | | 0.001 | 5.66 | 1200 | 0.0212 | 0.7723 | | 0.0011 | 6.6 | 1400 | 0.0217 | 0.7707 | | 0.0011 | 7.55 | 1600 | 0.0211 | 0.7724 | | 0.001 | 8.49 | 1800 | 0.0213 | 0.7716 | | 0.001 | 9.43 | 2000 | 0.0208 | 0.7702 | | 0.0009 | 10.38 | 2200 | 0.0213 | 0.7685 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/xlm-roberta-base_pan_corr_5e-06
Shijia
2024-02-17T02:25:46Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:25:02Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_pan_corr_5e-06 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_pan_corr_5e-06 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0244 - Spearman Corr: 0.7659 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.85 | 200 | 0.0258 | 0.7651 | | No log | 1.69 | 400 | 0.0241 | 0.7668 | | 0.0007 | 2.54 | 600 | 0.0254 | 0.7642 | | 0.0007 | 3.38 | 800 | 0.0249 | 0.7668 | | 0.0007 | 4.23 | 1000 | 0.0247 | 0.7614 | | 0.0007 | 5.07 | 1200 | 0.0234 | 0.7675 | | 0.0007 | 5.92 | 1400 | 0.0249 | 0.7660 | | 0.0007 | 6.77 | 1600 | 0.0241 | 0.7653 | | 0.0007 | 7.61 | 1800 | 0.0235 | 0.7663 | | 0.0007 | 8.46 | 2000 | 0.0255 | 0.7692 | | 0.0007 | 9.3 | 2200 | 0.0241 | 0.7680 | | 0.0013 | 10.15 | 2400 | 0.0239 | 0.7660 | | 0.0013 | 10.99 | 2600 | 0.0239 | 0.7655 | | 0.0013 | 11.84 | 2800 | 0.0241 | 0.7671 | | 0.0017 | 12.68 | 3000 | 0.0241 | 0.7633 | | 0.0017 | 13.53 | 3200 | 0.0249 | 0.7658 | | 0.0016 | 14.38 | 3400 | 0.0244 | 0.7673 | | 0.0016 | 15.22 | 3600 | 0.0244 | 0.7659 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/furina_esp_corr_2e-05
Shijia
2024-02-17T02:13:25Z
101
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:12:16Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_esp_corr_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_esp_corr_2e-05 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0212 - Spearman Corr: 0.7706 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.94 | 200 | 0.0219 | 0.7728 | | No log | 1.89 | 400 | 0.0216 | 0.7705 | | 0.0013 | 2.83 | 600 | 0.0212 | 0.7740 | | 0.0013 | 3.77 | 800 | 0.0234 | 0.7700 | | 0.0012 | 4.72 | 1000 | 0.0214 | 0.7691 | | 0.0012 | 5.66 | 1200 | 0.0212 | 0.7732 | | 0.0011 | 6.6 | 1400 | 0.0213 | 0.7725 | | 0.0011 | 7.55 | 1600 | 0.0211 | 0.7716 | | 0.001 | 8.49 | 1800 | 0.0210 | 0.7724 | | 0.001 | 9.43 | 2000 | 0.0207 | 0.7712 | | 0.0009 | 10.38 | 2200 | 0.0212 | 0.7706 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/xlm-roberta-base_kin_loss_2e-05
Shijia
2024-02-17T02:09:53Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:09:09Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_kin_loss_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_kin_loss_2e-05 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0212 - Spearman Corr: 0.7789 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.89 | 200 | 0.0204 | 0.7803 | | No log | 1.78 | 400 | 0.0212 | 0.7812 | | 0.0005 | 2.67 | 600 | 0.0218 | 0.7836 | | 0.0005 | 3.56 | 800 | 0.0212 | 0.7810 | | 0.0004 | 4.45 | 1000 | 0.0212 | 0.7799 | | 0.0004 | 5.35 | 1200 | 0.0214 | 0.7792 | | 0.0005 | 6.24 | 1400 | 0.0208 | 0.7774 | | 0.0005 | 7.13 | 1600 | 0.0210 | 0.7804 | | 0.0005 | 8.02 | 1800 | 0.0212 | 0.7789 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/xlm-roberta-base_kin_corr_2e-05
Shijia
2024-02-17T02:02:31Z
89
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T02:01:52Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_kin_corr_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_kin_corr_2e-05 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0213 - Spearman Corr: 0.7790 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.89 | 200 | 0.0210 | 0.7800 | | No log | 1.78 | 400 | 0.0210 | 0.7820 | | 0.0006 | 2.67 | 600 | 0.0208 | 0.7800 | | 0.0006 | 3.56 | 800 | 0.0209 | 0.7822 | | 0.0006 | 4.45 | 1000 | 0.0212 | 0.7801 | | 0.0006 | 5.35 | 1200 | 0.0214 | 0.7792 | | 0.0005 | 6.24 | 1400 | 0.0213 | 0.7768 | | 0.0005 | 7.13 | 1600 | 0.0211 | 0.7803 | | 0.0005 | 8.02 | 1800 | 0.0210 | 0.7785 | | 0.0005 | 8.91 | 2000 | 0.0213 | 0.7783 | | 0.0005 | 9.8 | 2200 | 0.0211 | 0.7809 | | 0.0005 | 10.69 | 2400 | 0.0213 | 0.7790 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/xlm-roberta-base_kin_corr_5e-06
Shijia
2024-02-17T02:00:39Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:59:58Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_kin_corr_5e-06 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_kin_corr_5e-06 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0247 - Spearman Corr: 0.7669 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.89 | 200 | 0.0246 | 0.7645 | | No log | 1.78 | 400 | 0.0248 | 0.7676 | | 0.002 | 2.67 | 600 | 0.0237 | 0.7655 | | 0.002 | 3.56 | 800 | 0.0243 | 0.7647 | | 0.0019 | 4.45 | 1000 | 0.0244 | 0.7623 | | 0.0019 | 5.35 | 1200 | 0.0233 | 0.7655 | | 0.0018 | 6.24 | 1400 | 0.0247 | 0.7667 | | 0.0018 | 7.13 | 1600 | 0.0244 | 0.7635 | | 0.0018 | 8.02 | 1800 | 0.0249 | 0.7648 | | 0.0018 | 8.91 | 2000 | 0.0247 | 0.7669 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/furina_eng_loss_2e-05
Shijia
2024-02-17T01:59:06Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:58:09Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_eng_loss_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_eng_loss_2e-05 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0202 - Spearman Corr: 0.7777 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 1.33 | 200 | 0.0211 | 0.7732 | | 0.0009 | 2.66 | 400 | 0.0215 | 0.7773 | | 0.0008 | 3.99 | 600 | 0.0209 | 0.7752 | | 0.0008 | 5.32 | 800 | 0.0197 | 0.7734 | | 0.0007 | 6.64 | 1000 | 0.0211 | 0.7735 | | 0.0006 | 7.97 | 1200 | 0.0208 | 0.7751 | | 0.0006 | 9.3 | 1400 | 0.0203 | 0.7789 | | 0.0008 | 10.63 | 1600 | 0.0200 | 0.7797 | | 0.001 | 11.96 | 1800 | 0.0207 | 0.7734 | | 0.001 | 13.29 | 2000 | 0.0203 | 0.7756 | | 0.0009 | 14.62 | 2200 | 0.0202 | 0.7745 | | 0.0009 | 15.95 | 2400 | 0.0202 | 0.7777 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
bartowski/KunoichiLake-2x7b-exl2
bartowski
2024-02-17T01:50:30Z
5
2
null
[ "text-generation", "license:apache-2.0", "region:us" ]
text-generation
2024-02-17T01:22:41Z
--- license: apache-2.0 quantized_by: bartowski pipeline_tag: text-generation --- ## Exllama v2 Quantizations of KunoichiLake-2x7b Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.0.13">turboderp's ExLlamaV2 v0.0.13</a> for quantization. <b>The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)</b> Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions. Original model: https://huggingface.co/macadeliccc/KunoichiLake-2x7b | Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description | | ----- | ---- | ------- | ------ | ------ | ------ | ------------ | | [8_0](https://huggingface.co/bartowski/KunoichiLake-2x7b-exl2/tree/8_0) | 8.0 | 8.0 | 13.7 GB | 15.1 GB | 17.2 GB | Maximum quality that ExLlamaV2 can produce, near unquantized performance. | | [6_5](https://huggingface.co/bartowski/KunoichiLake-2x7b-exl2/tree/6_5) | 6.5 | 8.0 | 11.5 GB | 12.9 GB | 15.0 GB | Near unquantized performance at vastly reduced size, **recommended**. | | [5_0](https://huggingface.co/bartowski/KunoichiLake-2x7b-exl2/tree/5_0) | 5.0 | 6.0 | 9.3 GB | 10.7 GB | 12.8 GB | Slightly lower quality vs 6.5, great for 12gb cards with 16k context. | | [4_25](https://huggingface.co/bartowski/KunoichiLake-2x7b-exl2/tree/4_25) | 4.25 | 6.0 | 8.2 GB | 9.6 GB | 11.7 GB | GPTQ equivalent bits per weight. | | [3_5](https://huggingface.co/bartowski/KunoichiLake-2x7b-exl2/tree/3_5) | 3.5 | 6.0 | 7.0 GB | 8.4 GB | 10.5 GB | Lower quality, not recommended. | ## Download instructions With git: ```shell git clone --single-branch --branch 6_5 https://huggingface.co/bartowski/KunoichiLake-2x7b-exl2 KunoichiLake-2x7b-exl2-6_5 ``` With huggingface hub (credit to TheBloke for instructions): ```shell pip3 install huggingface-hub ``` To download the `main` (only useful if you only care about measurement.json) branch to a folder called `KunoichiLake-2x7b-exl2`: ```shell mkdir KunoichiLake-2x7b-exl2 huggingface-cli download bartowski/KunoichiLake-2x7b-exl2 --local-dir KunoichiLake-2x7b-exl2 --local-dir-use-symlinks False ``` To download from a different branch, add the `--revision` parameter: Linux: ```shell mkdir KunoichiLake-2x7b-exl2-6_5 huggingface-cli download bartowski/KunoichiLake-2x7b-exl2 --revision 6_5 --local-dir KunoichiLake-2x7b-exl2-6_5 --local-dir-use-symlinks False ``` Windows (which apparently doesn't like _ in folders sometimes?): ```shell mkdir KunoichiLake-2x7b-exl2-6.5 huggingface-cli download bartowski/KunoichiLake-2x7b-exl2 --revision 6_5 --local-dir KunoichiLake-2x7b-exl2-6.5 --local-dir-use-symlinks False ``` Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski
Harmj0y/nemesis-reranker
Harmj0y
2024-02-17T01:47:02Z
165
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:44:54Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Den-sota/lab1_finetuning
Den-sota
2024-02-17T01:44:14Z
121
0
transformers
[ "transformers", "safetensors", "marian", "text2text-generation", "generated_from_trainer", "dataset:kde4", "base_model:Helsinki-NLP/opus-mt-en-fr", "base_model:finetune:Helsinki-NLP/opus-mt-en-fr", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-02-16T22:50:23Z
--- license: apache-2.0 base_model: Helsinki-NLP/opus-mt-en-fr tags: - generated_from_trainer datasets: - kde4 metrics: - bleu model-index: - name: lab1_finetuning results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: kde4 type: kde4 config: en-fr split: train args: en-fr metrics: - name: Bleu type: bleu value: 52.88398487672078 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # lab1_finetuning This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset. It achieves the following results on the evaluation set: - Loss: 0.8556 - Bleu: 52.8840 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/xlm-roberta-base_ind_corr_2e-05
Shijia
2024-02-17T01:44:14Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:43:28Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_ind_corr_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_ind_corr_2e-05 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0213 - Spearman Corr: 0.7807 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.85 | 200 | 0.0222 | 0.7803 | | No log | 1.69 | 400 | 0.0209 | 0.7788 | | 0.0006 | 2.54 | 600 | 0.0208 | 0.7742 | | 0.0006 | 3.38 | 800 | 0.0211 | 0.7780 | | 0.0005 | 4.23 | 1000 | 0.0212 | 0.7815 | | 0.0005 | 5.07 | 1200 | 0.0211 | 0.7822 | | 0.0005 | 5.92 | 1400 | 0.0207 | 0.7796 | | 0.0004 | 6.77 | 1600 | 0.0223 | 0.7791 | | 0.0004 | 7.61 | 1800 | 0.0219 | 0.7789 | | 0.0004 | 8.46 | 2000 | 0.0212 | 0.7802 | | 0.0004 | 9.3 | 2200 | 0.0213 | 0.7821 | | 0.0004 | 10.15 | 2400 | 0.0212 | 0.7789 | | 0.0004 | 10.99 | 2600 | 0.0213 | 0.7786 | | 0.0004 | 11.84 | 2800 | 0.0213 | 0.7807 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/furina_eng_corr_2e-05
Shijia
2024-02-17T01:43:08Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:42:17Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_eng_corr_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_eng_corr_2e-05 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0203 - Spearman Corr: 0.7758 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 1.33 | 200 | 0.0216 | 0.7729 | | 0.0014 | 2.66 | 400 | 0.0212 | 0.7735 | | 0.0013 | 3.99 | 600 | 0.0214 | 0.7754 | | 0.0013 | 5.32 | 800 | 0.0215 | 0.7733 | | 0.0012 | 6.64 | 1000 | 0.0211 | 0.7700 | | 0.0012 | 7.97 | 1200 | 0.0203 | 0.7745 | | 0.0012 | 9.3 | 1400 | 0.0204 | 0.7792 | | 0.0011 | 10.63 | 1600 | 0.0199 | 0.7773 | | 0.001 | 11.96 | 1800 | 0.0210 | 0.7735 | | 0.001 | 13.29 | 2000 | 0.0204 | 0.7755 | | 0.001 | 14.62 | 2200 | 0.0203 | 0.7734 | | 0.0009 | 15.95 | 2400 | 0.0206 | 0.7752 | | 0.0009 | 17.28 | 2600 | 0.0205 | 0.7729 | | 0.0009 | 18.6 | 2800 | 0.0208 | 0.7732 | | 0.0008 | 19.93 | 3000 | 0.0203 | 0.7758 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
mitchyAI/garammchy
mitchyAI
2024-02-17T01:39:26Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2024-02-17T01:37:35Z
--- license: creativeml-openrail-m ---
Shijia/xlm-roberta-base_ind_corr_5e-06
Shijia
2024-02-17T01:38:55Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:38:10Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_ind_corr_5e-06 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_ind_corr_5e-06 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0245 - Spearman Corr: 0.7641 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.85 | 200 | 0.0236 | 0.7650 | | No log | 1.69 | 400 | 0.0247 | 0.7661 | | 0.0011 | 2.54 | 600 | 0.0257 | 0.7631 | | 0.0011 | 3.38 | 800 | 0.0244 | 0.7624 | | 0.001 | 4.23 | 1000 | 0.0235 | 0.7617 | | 0.001 | 5.07 | 1200 | 0.0242 | 0.7668 | | 0.001 | 5.92 | 1400 | 0.0245 | 0.7645 | | 0.0011 | 6.77 | 1600 | 0.0242 | 0.7619 | | 0.0011 | 7.61 | 1800 | 0.0232 | 0.7671 | | 0.0013 | 8.46 | 2000 | 0.0257 | 0.7673 | | 0.0013 | 9.3 | 2200 | 0.0242 | 0.7675 | | 0.0019 | 10.15 | 2400 | 0.0243 | 0.7645 | | 0.0019 | 10.99 | 2600 | 0.0241 | 0.7643 | | 0.0019 | 11.84 | 2800 | 0.0246 | 0.7649 | | 0.0019 | 12.68 | 3000 | 0.0248 | 0.7617 | | 0.0019 | 13.53 | 3200 | 0.0250 | 0.7644 | | 0.0018 | 14.38 | 3400 | 0.0247 | 0.7659 | | 0.0018 | 15.22 | 3600 | 0.0249 | 0.7649 | | 0.0018 | 16.07 | 3800 | 0.0245 | 0.7641 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/furina_kin_loss_0.0001
Shijia
2024-02-17T01:36:49Z
100
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:35:54Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_kin_loss_0.0001 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_kin_loss_0.0001 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0473 - Spearman Corr: nan ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.89 | 200 | 0.0462 | nan | | No log | 1.78 | 400 | 0.0487 | nan | | 0.0481 | 2.67 | 600 | 0.0482 | nan | | 0.0481 | 3.56 | 800 | 0.0469 | nan | | 0.0482 | 4.45 | 1000 | 0.0481 | nan | | 0.0482 | 5.35 | 1200 | 0.0493 | nan | | 0.0481 | 6.24 | 1400 | 0.0467 | nan | | 0.0481 | 7.13 | 1600 | 0.0478 | nan | | 0.0483 | 8.02 | 1800 | 0.0473 | nan | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/xlm-roberta-base_hin_loss_2e-05
Shijia
2024-02-17T01:32:41Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:31:59Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_hin_loss_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_hin_loss_2e-05 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0212 - Spearman Corr: 0.7807 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.85 | 200 | 0.0238 | 0.7825 | | No log | 1.69 | 400 | 0.0213 | 0.7820 | | 0.0007 | 2.54 | 600 | 0.0209 | 0.7744 | | 0.0007 | 3.38 | 800 | 0.0208 | 0.7817 | | 0.0007 | 4.23 | 1000 | 0.0211 | 0.7791 | | 0.0007 | 5.07 | 1200 | 0.0207 | 0.7817 | | 0.0007 | 5.92 | 1400 | 0.0212 | 0.7833 | | 0.0005 | 6.77 | 1600 | 0.0210 | 0.7803 | | 0.0005 | 7.61 | 1800 | 0.0207 | 0.7794 | | 0.0004 | 8.46 | 2000 | 0.0212 | 0.7809 | | 0.0004 | 9.3 | 2200 | 0.0212 | 0.7807 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
jeiku/Cookie_7B
jeiku
2024-02-17T01:23:43Z
55
3
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "arxiv:2311.03099", "arxiv:2306.01708", "base_model:jeiku/Rainbow_69_7B", "base_model:merge:jeiku/Rainbow_69_7B", "base_model:jeiku/SpaghettiOs_7B", "base_model:merge:jeiku/SpaghettiOs_7B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-02-15T20:27:16Z
--- base_model: - jeiku/SpaghettiOs_7B - jeiku/Rainbow_69_7B library_name: transformers tags: - mergekit - merge license: other --- # Cookie A reasonably logical model with a few datasets thrown in to increase RP abilities. This is a good candidate for a balanced 7B model that can provide assistant functionality alongside roleplaying or romantic endeavors. ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/626dfb8786671a29c715f8a9/ihl6LZY3smChRokJkHm9Q.jpeg) This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [jeiku/SpaghettiOs_7B](https://huggingface.co/jeiku/SpaghettiOs_7B) as a base. ### Models Merged The following models were included in the merge: * [jeiku/Rainbow_69_7B](https://huggingface.co/jeiku/Rainbow_69_7B) ### Configuration The following YAML configuration was used to produce this model: ```yaml merge_method: dare_ties base_model: jeiku/SpaghettiOs_7B parameters: normalize: true models: - model: jeiku/SpaghettiOs_7B parameters: weight: 1 - model: jeiku/Rainbow_69_7B parameters: weight: 1 dtype: float16 ```
Shijia/xlm-roberta-base_hin_corr_2e-05
Shijia
2024-02-17T01:23:36Z
90
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:22:57Z
--- license: mit base_model: FacebookAI/xlm-roberta-base tags: - generated_from_trainer model-index: - name: xlm-roberta-base_hin_corr_2e-05 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base_hin_corr_2e-05 This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0209 - Spearman Corr: 0.7808 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.85 | 200 | 0.0208 | 0.7783 | | No log | 1.69 | 400 | 0.0214 | 0.7813 | | 0.0006 | 2.54 | 600 | 0.0214 | 0.7775 | | 0.0006 | 3.38 | 800 | 0.0211 | 0.7811 | | 0.0006 | 4.23 | 1000 | 0.0208 | 0.7799 | | 0.0006 | 5.07 | 1200 | 0.0213 | 0.7807 | | 0.0006 | 5.92 | 1400 | 0.0218 | 0.7775 | | 0.0006 | 6.77 | 1600 | 0.0206 | 0.7817 | | 0.0006 | 7.61 | 1800 | 0.0213 | 0.7821 | | 0.0005 | 8.46 | 2000 | 0.0213 | 0.7804 | | 0.0005 | 9.3 | 2200 | 0.0218 | 0.7812 | | 0.0004 | 10.15 | 2400 | 0.0215 | 0.7793 | | 0.0004 | 10.99 | 2600 | 0.0215 | 0.7794 | | 0.0004 | 11.84 | 2800 | 0.0212 | 0.7815 | | 0.0004 | 12.68 | 3000 | 0.0221 | 0.7763 | | 0.0004 | 13.53 | 3200 | 0.0215 | 0.7782 | | 0.0004 | 14.38 | 3400 | 0.0209 | 0.7808 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/furina_hau_corr_5e-06
Shijia
2024-02-17T01:21:35Z
89
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:20:42Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_hau_corr_5e-06 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_hau_corr_5e-06 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0216 - Spearman Corr: 0.7684 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.95 | 200 | 0.0214 | 0.7675 | | No log | 1.91 | 400 | 0.0218 | 0.7668 | | 0.0028 | 2.86 | 600 | 0.0223 | 0.7686 | | 0.0028 | 3.82 | 800 | 0.0231 | 0.7674 | | 0.0028 | 4.77 | 1000 | 0.0226 | 0.7669 | | 0.0028 | 5.73 | 1200 | 0.0216 | 0.7665 | | 0.0026 | 6.68 | 1400 | 0.0223 | 0.7697 | | 0.0026 | 7.64 | 1600 | 0.0216 | 0.7655 | | 0.0025 | 8.59 | 1800 | 0.0215 | 0.7686 | | 0.0025 | 9.55 | 2000 | 0.0216 | 0.7681 | | 0.0024 | 10.5 | 2200 | 0.0212 | 0.7672 | | 0.0024 | 11.46 | 2400 | 0.0219 | 0.7669 | | 0.0024 | 12.41 | 2600 | 0.0220 | 0.7657 | | 0.0024 | 13.37 | 2800 | 0.0217 | 0.7654 | | 0.0023 | 14.32 | 3000 | 0.0216 | 0.7684 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
wrchen1/testtesttest
wrchen1
2024-02-17T01:20:04Z
120
0
transformers
[ "transformers", "safetensors", "marian", "text2text-generation", "generated_from_trainer", "dataset:kde4", "base_model:Helsinki-NLP/opus-mt-en-fr", "base_model:finetune:Helsinki-NLP/opus-mt-en-fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-02-17T01:19:01Z
--- license: apache-2.0 base_model: Helsinki-NLP/opus-mt-en-fr tags: - generated_from_trainer datasets: - kde4 model-index: - name: testtesttest results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # testtesttest This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0 - Datasets 2.17.0 - Tokenizers 0.15.2
Shijia/furina_ind_loss_0.0001
Shijia
2024-02-17T01:18:53Z
89
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:yihongLiu/furina", "base_model:finetune:yihongLiu/furina", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-02-17T01:17:47Z
--- base_model: yihongLiu/furina tags: - generated_from_trainer model-index: - name: furina_ind_loss_0.0001 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # furina_ind_loss_0.0001 This model is a fine-tuned version of [yihongLiu/furina](https://huggingface.co/yihongLiu/furina) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0465 - Spearman Corr: nan ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 128 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Spearman Corr | |:-------------:|:-----:|:----:|:---------------:|:-------------:| | No log | 0.85 | 200 | 0.0466 | nan | | No log | 1.69 | 400 | 0.0475 | nan | | 0.047 | 2.54 | 600 | 0.0480 | nan | | 0.047 | 3.38 | 800 | 0.0469 | nan | | 0.0475 | 4.23 | 1000 | 0.0471 | nan | | 0.0475 | 5.07 | 1200 | 0.0474 | nan | | 0.0475 | 5.92 | 1400 | 0.0476 | nan | | 0.0476 | 6.77 | 1600 | 0.0476 | nan | | 0.0476 | 7.61 | 1800 | 0.0465 | nan | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.2
wrchen1/lab1_ran123
wrchen1
2024-02-17T01:10:12Z
119
0
transformers
[ "transformers", "safetensors", "marian", "text2text-generation", "generated_from_trainer", "dataset:kde4", "base_model:Helsinki-NLP/opus-mt-en-fr", "base_model:finetune:Helsinki-NLP/opus-mt-en-fr", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-02-17T01:07:29Z
--- license: apache-2.0 base_model: Helsinki-NLP/opus-mt-en-fr tags: - generated_from_trainer datasets: - kde4 model-index: - name: lab1_ran123 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # lab1_ran123 This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0 - Datasets 2.17.0 - Tokenizers 0.15.2