I’ve tried coding and every one I’ve tried fails unless really, really basic small functions like what you learn as a newbie compared to say 4o mini that can spit out more sensible stuff that works.

I’ve tried explanations and they just regurgitate sentences that can be irrelevant, wrong, or get stuck in a loop.

So. what can I actually use a small LLM for? Which ones? I ask because I have an old laptop and the GPU can’t really handle anything above 4B in a timely manner. 8B is about 1 t/s!

    • wise_pancake@lemmy.ca
      link
      fedilink
      English
      arrow-up
      1
      ·
      4 hours ago

      Open webui lets you install a ton of different search providers out of the box, but you do need sn API key for most and I haven’t vetted them

      I’m trying to get Kagi to work with Phi4 and not having success.

      • catty@lemmy.worldOP
        link
        fedilink
        English
        arrow-up
        1
        ·
        3 hours ago

        Thanks, when I get some time soon, I’ll have another look at it and cherry ai with a local install of ollama

    • herseycokguzelolacak@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      ·
      3 days ago

      Not on top of my head, but there must be something. llama.cpp and vllm have basically solved the inference problem for LLMs. What you need is a RAG solution on top that also combines it with web search.