A full finetune of Llama 2 7B using my Alpaca-transformed CoEdIT dataset. I gave it three epochs of training using a single A100 80GB GPU.
The intent was to create a L2 model that specializes in grammar correction. Results may vary.
Prompt Format
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
Remove all grammatical errors from this text: <insert text here>
### Response:
- Downloads last month
- 9
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.