File size: 1,828 Bytes
8839122 1b0fdde 5d487ff 8839122 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 | ---
base_model:
- meta-llama/Llama-3.1-8B-Instruct
tags:
- conversational
- roleplay
- chat
license: llama3.1
---
# Llama 3.1 8b RP Ink

A roleplay-focused LoRA finetune of Llama 3.1 8B Instruct. Methodology and hyperparams inspired by [SorcererLM](https://huggingface.co/rAIfle/SorcererLM-8x22b-bf16) and [Slush](https://huggingface.co/crestf411/Q2.5-32B-Slush).
Yet another model in the Ink series, following in the footsteps of [the rest of them](https://huggingface.co/collections/allura-org/ink-6772fd1442308781594bbabb)
## Dataset
The worst mix of data you've ever seen. Like, seriously, you do not want to see the things that went into this model. It's bad.
"this is like washing down an adderall with a bottle of methylated rotgut" - inflatebot
Update: I have sent the (public datasets in the) data mix publicly already so here's that
<details>
<img src=https://huggingface.co/proxy/cdn-uploads.huggingface.co/production/uploads/634262af8d8089ebaefd410e/JtjUoKtbOfBZfSSKojTcj.png>
</details>
## Quants
TODO!
## Recommended Settings
Chat template: Llama 3.1
Recommended samplers (not the be-all-end-all, try some on your own!):
- Temp 1.03 / Top A 0.3 / TFS 0.75 / Rep Pen 1.03
- Your samplers can go here! :3
## Hyperparams
### General
- Epochs = 2
- LR = 6e-5
- LR Scheduler = Cosine
- Optimizer = Paged AdamW 8bit
- Effective batch size = 16
### LoRA
- Rank = 16
- Alpha = 32
- Dropout = 0.25 (Inspiration: [Slush](https://huggingface.co/crestf411/Q2.5-32B-Slush))
## Credits
Humongous thanks to the people who created and curated the original data
Big thanks to all Allura members, for testing and emotional support ilya /platonic
especially to inflatebot who made the model card's image :3 |