Whisprell-DeepSeek-R1-Enhanced-1.5B

NexThinkLabs

Introduction

Whisprell-DeepSeek-R1-Enhanced-1.5B is a Chain-of-Thought (CoT) reasoning focused model developed by NexThinkLabs. The model is based on DeepSeek's DeepSeek-R1-Distill-Qwen-1.5B and has been further fine-tuned to enhance reasoning capabilities while maintaining computational efficiency.

Model Details

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("NexThinkLabsAI/Whisprell-DeepSeek-R1-Enhanced-1.5B)
tokenizer = AutoTokenizer.from_pretrained("NexThinkLabsAI/Whisprell-DeepSeek-R1-Enhanced-1.5B")

Usage Recommendations

  1. Temperature: 0.5-0.7 (0.6 recommended)
  2. Avoid system prompts - include instructions in user prompt
  3. For math problems: Include "Please reason step by step, and put your final answer within \boxed{}"
  4. Enforce thinking pattern with "<think>\n" at response start

License

This model is under Personal Proprietary License. The base model (deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) is under MIT License.

Acknowledgments

We thank DeepSeek AI for their DeepSeek-R1-Distill-Qwen-1.5B model which served as the foundation for this work.

Contact

For questions and support, please:

Downloads last month
0
Safetensors
Model size
1.78B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for NexThinkLabsAI/Whisprell-DeepSeek-R1-Enhanced-1.5B

Finetuned
(1)
this model

Collection including NexThinkLabsAI/Whisprell-DeepSeek-R1-Enhanced-1.5B