running on local machine
#19
by
saidavanam
- opened
I have laptop which has graphics gtx 1050 in it.My question is whether am i able to run this model on my machine locally or not??
no
Ask feom Deepseek chat and get a good guideline
@Dragonfly81 Thanks
One can run it locally by use of the llama model
so can we run on single 3090 TI. Or I have 3 x 3090 but connected with lan can a I distrubute with ray?
Another option that would work perfectly well, is to use the inteference to run the model. try lm studio