--- base_model: unsloth/qwen2.5-3b-instruct-unsloth-bnb-4bit tags: - text-generation-inference - transformers - unsloth - qwen2 - gguf - trl - grpo license: apache-2.0 language: - en --- # L'ia is a Reasoning Model trained on 500 steps in Unsloth GRPO based on 49,909 (oops) datapoints of Lambda Calculus. ## This active experiment is a test to see if a model can be trained to primarily think in pure simple logic. Maybe after more time we can make the Vulcans proud! ![Lia-creekside1.png](https://cdn-uploads.huggingface.co/production/uploads/65ebe4fba69aaabb4304def8/VV182sNXCq_uCFMSi1zbE.png) - **Developed by:** Creekside - **License:** apache-2.0 - **Finetuned from model :** unsloth/qwen2.5-3b-instruct-unsloth-bnb-4bit This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [](https://github.com/unslothai/unsloth)