Instructions to use fgaim/tielectra-small with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use fgaim/tielectra-small with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("fill-mask", model="fgaim/tielectra-small")# Load model directly from transformers import AutoTokenizer, AutoModelForMaskedLM tokenizer = AutoTokenizer.from_pretrained("fgaim/tielectra-small") model = AutoModelForMaskedLM.from_pretrained("fgaim/tielectra-small") - Notebooks
- Google Colab
- Kaggle
Pre-trained ELECTRA small for Tigrinya Language
We pre-train ELECTRA small on the TLMD dataset, with over 40 million tokens.
Contained are trained Flax and PyTorch models.
Hyperparameters
The hyperparameters corresponding to model sizes mentioned above are as follows:
| Model Size | L | AH | HS | FFN | P | Seq |
|---|---|---|---|---|---|---|
| SMALL | 12 | 4 | 256 | 1024 | 14M | 512 |
(L = number of layers; AH = number of attention heads; HS = hidden size; FFN = feedforward network dimension; P = number of parameters; Seq = maximum sequence length.)
Framework versions
- Transformers 4.12.0.dev0
- Pytorch 1.9.0+cu111
- Datasets 1.13.3
- Tokenizers 0.10.3
Citation
If you use this model in your product or research, please cite as follows:
@article{Fitsum2021TiPLMs,
author={Fitsum Gaim and Wonsuk Yang and Jong C. Park},
title={Monolingual Pre-trained Language Models for Tigrinya},
year=2021,
publisher={WiNLP 2021 at EMNLP 2021}
}
- Downloads last month
- 93