It's been a bit of a given in the LLM world that there will be somewhat competitive open weights models. I'm not sure that's a good assumption anymore. A short history of LLMs In the relatively brief history of LLMs, there's been two[1] types of LLMs - closed and "open weights". Closed models include nearly everything from OpenAI (despite the name!) with open weights models being released from other labs. Famously the Llama series of models were open weights, but more recently the Chinese labs such as MiniMax, Z.ai, DeepSeek and Qwen (Alibaba) have been the leading open weights models, with Google's Gemma series and OpenAI's gpt-oss models generally coming somewhere behind the Chinese ones. Open weights models allow anyone to run the model on their own hardware. Typically models that were worth running required very beefy hardware - but this is rapidly changing, with smaller models becoming far more useful. Being able to run these models locally - as opposed to via an API to an…
No comments yet. Log in to reply on the Fediverse. Comments will appear here.