Comment by BloondAndDoom
11 hours ago
This pretty cool, and useful but I only wish this was a website. I don’t like the idea of running an executable for something that can perfectly be done as a website. (Other than some minor features, tbh even you can enable Corsair and still check the installed models from a web browser).
Sounds like a fun personal project though.
>I only wish this was a website. I don’t like the idea of running an executable for something that can perfectly be done as a website.
The tool depends on hardware detection. From https://github.com/AlexsJones/llmfit?tab=readme-ov-file#how-... :
Therefore, a website running Javascript is restricted by the browser sandbox so can't see the same low-level details such as total system RAM, exact count of GPUs, etc,
To implement your idea so it's only a website and also workaround the Javascript limitations, a different kind of workflow would be needed. E.g. run macOS system report to generate a .spx file, or run Linux inxi to generate a hardware devices report... and then upload those to the website for analysis to derive a "LLM best fit". But those os report files may still be missing some details that the github tool gathers.
Another way is to have the website with a bunch of hardware options where the user has to manually select the combination. Less convenient but then again, it has the advantage of doing "what-if" scenarios for hardware the user doesn't actually have and is thinking of buying.
(To be clear, I'm not endorsing this particular github tool. Just pointing out that a LLMfit website has technical limitations.)
That’s like like 4 or 5 fields to fill in on a form. Way less intrusive than installing this thing
It can become complicated when you run it inside a container.
3 replies →
Came across a website for this recently that may be worth a look https://whatmodelscanirun.com
It's wildly inaccurate for me.
I just discovered the other day the hugging face allows you to do exactly this.
With the caveat that you enter your hardware manually. But are we really at the point yet where people are running local models without knowing what they are running them on..?
The whole point is to measure your hardware capability. How would you do that as a website?
here's an website for a community-ran db on LLM models with details on configs for their token/s: https://inferbench.com/
Huggingface has it built in.
Where?
In your preferences there is a local apps and hardware, I guess it's a little different because I just open the page of a model and it shows the hardware I've configured and shows me what quants fit.
always liked this website that kinda does something similar https://apxml.com/tools/vram-calculator