What is everyone using for the LLM model for HA voice when selfhosting ollama? I’ve tried llama and qwen with varying degrees of understanding my commands. I’m currently on llama as it appears a little better. I just wanted to see if anyone found a better model.

Edit: as pointed out, this is more of a speech to text issue than llm model. I’m looking into the alternatives to whisper

    • spitfire@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      9 days ago

      So basically for people who have graphic cards with 24GB VRAM (or more). While I do, it’s probably something most people don’t ;)

        • spitfire@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          9 days ago

          I could probably run something on my gaming PC with 3090, but that would be a big cost. Instead I’ve just put my old 2070 in an existing server and using it for more lightweight stuff (TTS, obico, frigate, ollama with some small model).