ah, yeah, that's about 50% more grunty than my RX 7800 XT. it has 256 bit memory and 16gb. runs 22b codestral fine though.
these free models from hugging face are quite a jumble of hit and miss tho. took me a while to find a good one, and then someone put me onto codestral, which also seems to be quite good, and more parameters than the 14B qwen 3 i was using before. haven't really evaluated it though, because most of the work i do uses claude 3.7 cloud for a coding agent. i'm looking forward to eventually being able to point the agent at my local LLMs tho. i just don't see the point in using a remote service. i also don't like teaching those fuckers my process, that's more what i'm concerned about because i know copilot is already eating all of my output on github.