YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Prerequisites

After cloning the repository, first fetch submodule dependencies and run:

git submodule update --init --recursive

A Universal Dependency parser built on top of a Transformer language model

Python3.8 recommended, as well as a virtual environment.

You can use conda for a virtual environment: https://conda.io/projects/conda/en/latest/user-guide/getting-started.html You can also use venv for a virtual environment: https://docs.python.org/3/library/venv.html

To run this package, after having activated your virtual environment, you need to install the requirements: python3 -m pip install -r requirements.txt.

The Tokenizer submodule is using Miðeind's tokenizer. It is included because one of Diaparser's modules is named tokenizer.

The parser can be run as follows:

python3 parse_file.py --parser diaparser-is-combined-v211/diaparser.model --infile test_file.txt

The directory transformer_models/ contains a pretrained model, electra-base-igc-is, which supplies the parser with contextual embeddings and attention, trained by Jón Friðrik Daðason.

The parser scores as follows:

Metric     | Precision |    Recall |  F1 Score | AligndAcc
-----------+-----------+-----------+-----------+-----------
Tokens     |     99.70 |     99.77 |     99.73 |
Sentences  |    100.00 |    100.00 |    100.00 |
Words      |     99.62 |     99.61 |     99.61 |
UAS        |     89.58 |     89.57 |     89.58 |     89.92
LAS        |     86.46 |     86.45 |     86.46 |     86.79
CLAS       |     82.30 |     81.81 |     82.05 |     82.24

License

https://opensource.org/licenses/Apache-2.0

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.