LFM2.5-1.2B Italian

LiquidAI/LFM2.5-1.2B-Base model further continued-pretrained on Italian Wikipedia and then fine-tuned with Italian instructions (Alpaca format).

Training

  • Continued Pretraining: ~1% of Italian Wikipedia (20231101.it dump), texts ≥ 600 characters
  • Instruction Tuning: dataset DanielSc4/alpaca-cleaned-italian
  • Technique: LoRA + Unsloth (merged in full precision / fp16)
  • Original base model: multilingual, with significant improvement in Italian

Supported Languages

Improved in Italian.
Maintains English and residual capability in the other languages of the base model.

This model follows the exact continued pretraining + instruction tuning recipe recently published by LiquidAI in their official cookbook: https://github.com/Liquid4All/cookbook/blob/main/finetuning/notebooks/cpt_translation_with_unsloth.ipynb

Korean was replaced with Italian by using: Italian Wikipedia (20231101.it dump) instead of Korean Wikipedia for continued pretraining and DanielSc4/alpaca-cleaned-italian dataset instead of the Korean Alpaca for instruction tuning All other steps (LoRA including embed_tokens & lm_head, separate embedding learning rate, Unsloth, fp16 merge) are identical to the original notebook.

Downloads last month
26
Safetensors
Model size
1B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 1 Ask for provider support

Model tree for harrier77/LFM2.5-1.2B-ITA

Finetuned
(21)
this model

Evaluation results

  • Perplexity (not evaluated) on Italian Wikipedia (20231101.it) - continued pretraining
    self-reported
    null