Dataset used: mpasila/Literotica-stories-short which contains only a subset of the stories from the full Literotica dataset and was chunked down to fit within 8192 tokens.

Prompt format is: No formatting

Merged model: mpasila/Llama-3.1-Literotica-8B

Trained with regular LoRA (not quantized/QLoRA) and LoRA rank was 128 and Alpha set to 32. Trained for 1 epoch using A40 for about 13 hours.

Uploaded model

  • Developed by: mpasila
  • License: Llama 3.1 Community License Agreement
  • Finetuned from model : unsloth/meta-llama-3.1-8b

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
39
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for mpasila/Llama-3.1-Literotica-LoRA-8B

Merges
6 models

Dataset used to train mpasila/Llama-3.1-Literotica-LoRA-8B