Ask HN: What happened to self-hosted models?

3作者: curiousaboutml28 天前原帖
Hi HN, sorry for using a burner account.<p>It seems to me that up until the beginning of the last year, we saw a couple of new &quot;open&quot; model release announcements almost every week. They&#x27;d set a new state of the art for what an enthusiast could run on their laptop or home server.<p>Meta, Deepseek, Mistral, Qwen, even Google etc. were publishing new models left and right. There were new formats, quantizations, inference engines etc. and most importantly - a lot of discourse and excitement around them.<p>Quietly and suddenly, this changed. After the release of gpt-oss (August 2025), the discourse has been heavily dominated around hosted models now. I don&#x27;t think I&#x27;ve seen any mention of Ollama in any discussion that reached HN&#x27;s front page in the last 6 months.<p>What gives? Is this a proxy signal that we&#x27;ve hit a barrier in LLM efficiency?
查看原文
Hi HN, sorry for using a burner account.<p>It seems to me that up until the beginning of the last year, we saw a couple of new &quot;open&quot; model release announcements almost every week. They&#x27;d set a new state of the art for what an enthusiast could run on their laptop or home server.<p>Meta, Deepseek, Mistral, Qwen, even Google etc. were publishing new models left and right. There were new formats, quantizations, inference engines etc. and most importantly - a lot of discourse and excitement around them.<p>Quietly and suddenly, this changed. After the release of gpt-oss (August 2025), the discourse has been heavily dominated around hosted models now. I don&#x27;t think I&#x27;ve seen any mention of Ollama in any discussion that reached HN&#x27;s front page in the last 6 months.<p>What gives? Is this a proxy signal that we&#x27;ve hit a barrier in LLM efficiency?