Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Darkly)
  • No Skin
Collapse
Brand Logo
  1. Home
  2. Uncategorized
  3. Today I'm playing with my Raspberry Pi 5 16GB, and the new-fangled "AI Hat+ 2".

Today I'm playing with my Raspberry Pi 5 16GB, and the new-fangled "AI Hat+ 2".

Scheduled Pinned Locked Moved Uncategorized
raspberrypilinuxdockerollama
8 Posts 3 Posters 31 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • Owl EyesD This user is from outside of this forum
    Owl EyesD This user is from outside of this forum
    Owl Eyes
    wrote last edited by
    #1

    Today I'm playing with my Raspberry Pi 5 16GB, and the new-fangled "AI Hat+ 2". I don't really like docker, but will play along, following the documentation provided, to get a nice "Open WebUI" web interface for the chats:

    https://www.raspberrypi.com/documentation/computers/ai.html#step1-llm

    I'm curious if any of the models they provide are any good:
    - "deepseek_r1_distill_qwen:1.5b"
    - "llama3.2:3b"
    - "qwen2.5-coder:1.5b"
    - "qwen2.5-instruct:1.5b"
    - "qwen2:1.5b"

    Can anyone vouch for these models?

    #RaspberryPi #Linux #AI #docker #ollama #OpenSource

    HenriS AnthropyA 2 Replies Last reply
    0
    • Owl EyesD Owl Eyes

      Today I'm playing with my Raspberry Pi 5 16GB, and the new-fangled "AI Hat+ 2". I don't really like docker, but will play along, following the documentation provided, to get a nice "Open WebUI" web interface for the chats:

      https://www.raspberrypi.com/documentation/computers/ai.html#step1-llm

      I'm curious if any of the models they provide are any good:
      - "deepseek_r1_distill_qwen:1.5b"
      - "llama3.2:3b"
      - "qwen2.5-coder:1.5b"
      - "qwen2.5-instruct:1.5b"
      - "qwen2:1.5b"

      Can anyone vouch for these models?

      #RaspberryPi #Linux #AI #docker #ollama #OpenSource

      HenriS This user is from outside of this forum
      HenriS This user is from outside of this forum
      Henri
      wrote last edited by
      #2

      @d1 they are basically toy models

      Owl EyesD 1 Reply Last reply
      0
      • HenriS Henri

        @d1 they are basically toy models

        Owl EyesD This user is from outside of this forum
        Owl EyesD This user is from outside of this forum
        Owl Eyes
        wrote last edited by
        #3

        @slyecho woe is me, who spent too much on this toy

        HenriS 1 Reply Last reply
        0
        • Owl EyesD Owl Eyes

          @slyecho woe is me, who spent too much on this toy

          HenriS This user is from outside of this forum
          HenriS This user is from outside of this forum
          Henri
          wrote last edited by
          #4

          @d1 well, nothing wrong with toys. I also have a lot of SBCs and Raspberry Pi stuff to play with. But yeah, these models can run on the CPU too probably

          Owl EyesD 1 Reply Last reply
          0
          • HenriS Henri

            @d1 well, nothing wrong with toys. I also have a lot of SBCs and Raspberry Pi stuff to play with. But yeah, these models can run on the CPU too probably

            Owl EyesD This user is from outside of this forum
            Owl EyesD This user is from outside of this forum
            Owl Eyes
            wrote last edited by
            #5

            @slyecho I was hoping that the "whisper" audio-to-text LLMs would gain hardware acceleration for that Hailo NPU (on that AI Hat+ 2), but alas, there's no mention of Hailo/Raspberry Pi hardware yet, here:

            https://github.com/ggml-org/whisper.cpp

            #RaspberryPi #LLM #AI #whisper

            1 Reply Last reply
            0
            • Owl EyesD Owl Eyes

              Today I'm playing with my Raspberry Pi 5 16GB, and the new-fangled "AI Hat+ 2". I don't really like docker, but will play along, following the documentation provided, to get a nice "Open WebUI" web interface for the chats:

              https://www.raspberrypi.com/documentation/computers/ai.html#step1-llm

              I'm curious if any of the models they provide are any good:
              - "deepseek_r1_distill_qwen:1.5b"
              - "llama3.2:3b"
              - "qwen2.5-coder:1.5b"
              - "qwen2.5-instruct:1.5b"
              - "qwen2:1.5b"

              Can anyone vouch for these models?

              #RaspberryPi #Linux #AI #docker #ollama #OpenSource

              AnthropyA This user is from outside of this forum
              AnthropyA This user is from outside of this forum
              Anthropy
              wrote last edited by
              #6

              @d1 tiny models are fun to play with and can do basic things, but I'd hardly qualify a 1.5-3b parameter model as a "large" language model. It starts getting interesting at like 30-70b+, or 400b+ if you have the VRAM for it. Stuff like ChatGPT5.2 and Gemini3 Pro and such are trillions and trillions of parameters. Gemini 3 Flash is ~30b ish if I recall correctly.

              That said, you should definitely try play with them, they can still do basic things in small contexts.

              Owl EyesD 1 Reply Last reply
              2
              0
              • R AodeRelay shared this topic
                R ActivityRelay shared this topic
              • AnthropyA Anthropy

                @d1 tiny models are fun to play with and can do basic things, but I'd hardly qualify a 1.5-3b parameter model as a "large" language model. It starts getting interesting at like 30-70b+, or 400b+ if you have the VRAM for it. Stuff like ChatGPT5.2 and Gemini3 Pro and such are trillions and trillions of parameters. Gemini 3 Flash is ~30b ish if I recall correctly.

                That said, you should definitely try play with them, they can still do basic things in small contexts.

                Owl EyesD This user is from outside of this forum
                Owl EyesD This user is from outside of this forum
                Owl Eyes
                wrote last edited by
                #7

                @anthropy thanks for the consolation

                AnthropyA 1 Reply Last reply
                0
                • Owl EyesD Owl Eyes

                  @anthropy thanks for the consolation

                  AnthropyA This user is from outside of this forum
                  AnthropyA This user is from outside of this forum
                  Anthropy
                  wrote last edited by
                  #8

                  @d1 IIRC people use these kinda models for things like Home Assistant automation, so you can tell a voice assistant to turn on/off lights and such. In tiny contexts like that they seem to work fine- although I personally haven't tried that yet.

                  1 Reply Last reply
                  2
                  0
                  Reply
                  • Reply as topic
                  Log in to reply
                  • Oldest to Newest
                  • Newest to Oldest
                  • Most Votes


                  • Login

                  • Don't have an account? Register

                  • Login or register to search.
                  Powered by NodeBB Contributors
                  • First post
                    Last post
                  0
                  • Categories
                  • Recent
                  • Tags
                  • Popular
                  • World
                  • Users
                  • Groups