base model
#20 opened about 16 hours ago
by
ehartford
Can you upload your model to Ollama? I hope to use your model to run RAGflow. However, if I need to run it locally, it must be compatible with Ollama.
#19 opened 1 day ago
by
shaddock
strange messsage from using the model
#18 opened 4 days ago
by
lucas202
is it possible to build a rag using this model
1
#17 opened 4 days ago
by
lucas202
model weight
1
#16 opened 9 days ago
by
kdaeho27
Lenght Ouput ?
1
#15 opened 10 days ago
by
Brabuslevrai
Are there plans to release the lightning attention kernel?
2
#14 opened 11 days ago
by
bongchoi
In modeling_minimax_text_01.py attention mask is not passed correctly to MiniMaxText01FlashAttention2::forward() method
1
#13 opened 11 days ago
by
sszymczyk
Request: Add vLLM Support for This Model
1
#12 opened 13 days ago
by
kira
Can you provide a FP8 version?
2
#11 opened 13 days ago
by
xjpang85
Smaller versions (like 20b and 14b)
1
#10 opened 14 days ago
by
win10
Please fire your human evaluators
8
#6 opened 15 days ago
by
ChuckMcSneed
Consider making Minimax Text free software, as license is proprietary
4
#2 opened 16 days ago
by
JLouisBiz
Requesting Support for GGUF Quantization of MiniMax-Text-01 through llama.cpp
4
#1 opened 16 days ago
by
Doctor-Chad-PhD