Plus Light
Light4Bear
AI & ML interests
None yet
Recent Activity
new activity
19 days ago
cognitivecomputations/DeepSeek-R1-AWQ:Can't get 48 TPS on 8x H800
new activity
about 1 month ago
cognitivecomputations/DeepSeek-V3-AWQ:GPTQ Support
new activity
8 months ago
deepseek-ai/DeepSeek-V2-Chat:Join LMSYS Chatbot Arena?
Organizations
Light4Bear's activity
Can't get 48 TPS on 8x H800
1
#21 opened 19 days ago
by
Light4Bear

GPTQ Support
2
#1 opened 2 months ago
by
warlock-edward
Join LMSYS Chatbot Arena?
1
#11 opened 10 months ago
by
Light4Bear

Can VLLM be used for loading?
6
#4 opened 11 months ago
by
wawoshashi
How many bits and what is the groupsize?
1
#3 opened 11 months ago
by
vitvit
How to load command r+ in text-generation-webui?
5
#1 opened 11 months ago
by
MLDataScientist
GPTQ/AWQ quant that is runable in vllm?
2
#4 opened about 1 year ago
by
Light4Bear

Difference between v0.2 and v0.4?
1
#2 opened about 1 year ago
by
Light4Bear

Tokenizer is incorrectly tokenizing '<|im_start|>' and '<|im_end|>' as strings
#5 opened about 1 year ago
by
Light4Bear

Instruct sequences?
12
#2 opened about 1 year ago
by
deleted
What is the context size of this model?
8
#1 opened about 1 year ago
by
MarinaraSpaghetti

Suggestion: increase the weights of instruction-following models.
3
#6 opened about 1 year ago
by
Light4Bear

Safetensor naming convention
10
#1 opened about 1 year ago
by
dannysemi
vLLM output gibberish but text-generation-webui is fine
1
#2 opened about 1 year ago
by
Light4Bear

GPTQ and AWQ quants
1
#1 opened about 1 year ago
by
Light4Bear

Congratulations!
10
#1 opened about 1 year ago
by
TomGrc
Difference between this and the non-"test" version?
16
#2 opened about 1 year ago
by
samgreen

8.0bpw-h8-exl2 quant of this model
6
#1 opened about 1 year ago
by
Light4Bear

fp16 version of the model
1
#6 opened almost 2 years ago
by
Light4Bear
