Qwen3-0.6B is about 400 MB at Q4 and is surprisingly coherent for what it is.
yellow [she/her]
- 0 Posts
- 4 Comments
Joined 1 year ago
Cake day: November 12th, 2024
You are not logged in. If you use a Fediverse account that is able to follow users, you can follow this user.
yellow [she/her]@lemmy.blahaj.zoneto
Linux@programming.dev•Bazzite just delivered over a petabyte of ISOs in a single monthEnglish
4·4 hours agoI have to ask, why buy a key for the VM at all? Windows functions perfectly fine without one, and you can always use MAS if you want to change your wallpaper or something.
yellow [she/her]@lemmy.blahaj.zonetoMicroblog Memes@lemmy.world•words to live by 🙏English
2·23 hours agoGames should have a point, and winning is not a point on its own.
Why not? Is wanting to win not a valid motivator to play a game?

It’s not the LLM that does the web searching, but the software stack around it. On its own, an LLM is just a text completer. What you’d need a frontend like OpenWebUI or Perplexica that would ask the LLM for, say five internet search queries that could return useful information for the prompt, throw those queries into SearxNG, and then pipe the results into the LLM’s context for it to be used.
As for the models themselves, any decently-sized one that was released fairly recently would work. If you’re looking specifically for open-source rather than open-weight models (meaning that the training data and methodologies were also released rather than just the model weights), GPT-OSS 20B/120B and the OLMo models are recent standouts there. If not, the Qwen3 series are pretty good. (There are other good models out there, this is just what I remember off the top of my head.)