Cypriot-Greek BERT Model πŸ‡¨πŸ‡Ύ

A specialized BERT model fine-tuned for Masked Language Modeling (MLM) on Cypriot dialect and Modern Greek language pairs.

Model Details

  • Base Model: bert-base-greek-uncased-v1
  • Model Type: Masked Language Model (MLM)
  • Languages: Cypriot Greek (dialect), Modern Greek (standard)
  • Dataset Size: 30,000 Cypriot-Greek language pairs
  • Training Task: Bidirectional masked language modeling

Training Configuration

Hyperparameters

  • Learning Rate: 5e-5
  • Batch Size: 16 per device
  • Gradient Accumulation: 1 step
  • Epochs: 8
  • Warmup Steps: 1,000
  • Weight Decay: 0.01

Hardware & Optimization

  • Precision: bfloat16 (bf16)
  • Gradient Checkpointing: Disabled
  • Memory Optimization: Pin memory enabled
  • Data Loading: 4 workers for parallel processing
  • Hardware: A100 GPU 40gb

Steps

Step Train Loss Eval Loss
500 1.90 1.8217
750 1.9252 1.6843
1000 1.7017 1.6174
1500 1.3602 1.5243
2200 0.95 1.42
Downloads last month
7
Safetensors
Model size
113M params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Elormiden/bert-base-cypriot-greek

Finetuned
(8)
this model

Dataset used to train Elormiden/bert-base-cypriot-greek