为什么你该停止使用 Ollama - V2EX
V2EX = way to explore
V2EX 是一个关于分享和探索的地方
现在注册
已注册用户请  登录
catazshadow

为什么你该停止使用 Ollama

  •  1
     
  •   catazshadow 4 天前 3833 次点击
    原文:
    https://sleepingrobots.com/dreams/stop-using-ollama/

    “Ollama wrapped that work in a nice CLI, raised VC money on the back of it, spent over a year refusing to credit it, forked it badly, shipped a closed-source app alongside it, and then pivoted the whole thing toward cloud services. At every decision point where they could have been good open-source citizens, they chose the path that made them look more self-sufficient to investors.”

    总之就是开源小偷,还尝试锁死用户

    原文建议用这些:

    llama.cpp is the engine. It has an OpenAI-compatible API server (llama-server), a built-in web UI, full control over context windows and sampling parameters, and consistently better throughput than Ollama. In February 2026, Gerganov’s ggml.ai joined Hugging Face to ensure the long-term sustainability of the project. It’s truly community-driven, MIT-licensed, and under active development with 450+ contributors.

    llama-swap handles multi-model orchestration, loading, unloading, and hot-swapping models on demand behind a single API endpoint. Pair it with LiteLLM and you get a unified OpenAI-compatible proxy that routes across multiple backends with proper model aliasing.

    LM Studio gives you a GUI if that’s what you want. It uses llama.cpp under the hood, exposes all the knobs, and supports any GGUF model without lock-in. Jan is another open-source desktop app with a clean chat interface and local-first design. Msty offers a polished GUI with multi-model support and built-in RAG. koboldcpp is another option with a web UI and extensive configuration options.

    Red Hat’s ramalama is worth a look too, a container-native model runner that explicitly credits its upstream dependencies front and center. Exactly what Ollama should have done from the start.
    第 1 条附言    4 天前
    还有 ollama 的性能更差
    14 条回复    2026-04-22 11:36:19 +08:00
    anbabubabiluya
        1
    anbabubabiluya  
       3 天前 via Android
    有大佬能推荐一个部署平台吗?我也觉得 ollama 太慢了,显卡是 5060ti 16g ,最好能直接在 Windows 跑
    tool2dx
        2
    tool2dx  
       3 天前   1
    @anbabubabiluya ollama 不慢的, 我显卡比你还差, 只有 12G 显存, 但是电脑是双显卡, 加起来就有 24G 显存. 运行 ollama 上的 qwen3.6 35b-q4 版本, 如果优化后没爆显存, 速度满速飞起. 默认是爆显存 8%, 速度降为 1/6, 超慢.
    catazshadow
        3
    catazshadow  
    OP
       3 天前
    @anbabubabiluya lm studio 似乎可以
    ebushicao
        4
    ebushicao  
       3 天前
    我前段时间从 ollama 换到 lm studio 了,真的好很多,相比之下 ollama 真的是个一般的玩具
    r6cb
        5
    r6cb  
       3 天前
    @anbabubabiluya #1 试试看 wsl 安装 vllm
    woctordho
        6
    woctordho  
       3 天前 via Android
    @anbabubabiluya 用 llama.cpp 就行了
    metalvest
        7
    metalvest  
       3 天前 via Android
    对普通用户来说这些都可以忽略,好用就行,看看豆包为什么用户这么多就知道了
    rammiah
        8
    rammiah  
       3 天前
    lm-studio 支持服务器运行吗? ollama 主要是拉模型方便,modelscope ,hf 都能用
    01802
        9
    01802  
       3 天前 via Android
    最近不用 ollama ,为了方便,koboldcpp 都行
    catazshadow
        10
    catazshadow  
    OP
       3 天前 via Android
    @rammiah 服务器用 llama-swap 调 llama.cpp 就行了
    julyclyde
        11
    julyclyde  
       3 天前
    @rammiah 我总感觉 ollama 和 docker 有点像
    my3157
        12
    my3157  
       1 天前
    ollama ggml models 要切回 llama-server 了 https://github.com/ollama/ollama/pull/15122
    jackqian
        13
    jackqian  
       1 天前
    苹果用 omlx 是不是好点?
    seakingii
        14
    seakingii  
       5 小时 36 分钟前
    @rammiah 支持的.LM STUDIO 更方便.
    关于     帮助文档     自助推广系统     博客     API     FAQ     Solana     5461 人在线   最高记录 6679       Select Language
    创意工作者们的社区
    World is powered by solitude
    VERSION: 3.9.8.5 68ms UTC 09:13 PVG 17:13 LAX 02:13 JFK 05:13
    Do have faith in what you're doing.
    ubao msn snddm index pchome yahoo rakuten mypaper meadowduck bidyahoo youbao zxmzxm asda bnvcg cvbfg dfscv mmhjk xxddc yybgb zznbn ccubao uaitu acv GXCV ET GDG YH FG BCVB FJFH CBRE CBC GDG ET54 WRWR RWER WREW WRWER RWER SDG EW SF DSFSF fbbs ubao fhd dfg ewr dg df ewwr ewwr et ruyut utut dfg fgd gdfgt etg dfgt dfgd ert4 gd fgg wr 235 wer3 we vsdf sdf gdf ert xcv sdf rwer hfd dfg cvb rwf afb dfh jgh bmn lgh rty gfds cxv xcv xcs vdas fdf fgd cv sdf tert sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf shasha9178 shasha9178 shasha9178 shasha9178 shasha9178 liflif2 liflif2 liflif2 liflif2 liflif2 liblib3 liblib3 liblib3 liblib3 liblib3 zhazha444 zhazha444 zhazha444 zhazha444 zhazha444 dende5 dende denden denden2 denden21 fenfen9 fenf619 fen619 fenfe9 fe619 sdf sdf sdf sdf sdf zhazh90 zhazh0 zhaa50 zha90 zh590 zho zhoz zhozh zhozho zhozho2 lislis lls95 lili95 lils5 liss9 sdf0ty987 sdft876 sdft9876 sdf09876 sd0t9876 sdf0ty98 sdf0976 sdf0ty986 sdf0ty96 sdf0t76 sdf0876 df0ty98 sf0t876 sd0ty76 sdy76 sdf76 sdf0t76 sdf0ty9 sdf0ty98 sdf0ty987 sdf0ty98 sdf6676 sdf876 sd876 sd876 sdf6 sdf6 sdf9876 sdf0t sdf06 sdf0ty9776 sdf0ty9776 sdf0ty76 sdf8876 sdf0t sd6 sdf06 s688876 sd688 sdf86