What should one consider when choosing a model in #Huggingface?
Just the number of parameters and the filesize (the bigger these two are the better the model is)?
Which would be the best model in there? I tried running some models locally but they’re so much worse (“dumber”) than ChatGPT, I wanted to run something closer to it. Maybe the models I chose just weren’t big enough.
What is the “smarter”/biggest model in there?
I have an empty ssd I could use if the model is really big, and I have a good gpu so that’s not a problem either as long as the model doesn’t need like 10 gpus to run
Edit: okay apparently some models need like 300GB of VRAM to run, so lemme ask this differently:
What is the biggest, best model I can run on a RTX 4090? (24GB VRAM)
(64GB RAM)
Edit 2: apparently there’s this bitsandbytes thing that could help?
(Probably irrelevant but I’ll use the #oobabooga webUI)
#AI #LLM #GPT #GPT4all