Raspberry Pi's New AI Hat Adds 8GB of RAM for Local LLMs

(jeffgeerling.com)

45 points | by ingve 1 hour ago

15 comments

  • buran77 1 hour ago
    I think Raspberry lost the magic of the older Pis, they lost that sense of purpose. They basically created a niche with the first Pis, now they're just jumping into segments that others created and are already filled to the brim with perhaps even more qualified competition.

    Are they seeing a worthwhile niche for the tinkerers (or businesses?) who want to run local LLMs with middling performance but still need full set of GPIOs in a small package? Maybe. But maybe this is just Raspberry jumping on the bandwagon.

    I don't blame them for looking to expand into new segments, the business needs to survive. But these efforts just look a bit aimless to me. I "blame" them for not having another "Raspberry Pi moment".

    • joe_mamba 1 hour ago
      Yep. RPi foundation lost the plot a long time ago. The original RPi was in a league of its own when it launched since nothing like it existed and it was cheap.

      But now if I want some low power linux PC replacement with display output, for the price of the latest RPi 5, I can buy on the used market a ~2018 laptop with a 15W quad core CPU, 8GB RAM, 256 NVME and 1080p IPS display, that's orders of magnitude more capable. And if I want a battery powered embedded ARM device for GPIO over WIFI, I can get an ESP32 clone, that's orders of magnitude cheaper.

      Now RPi at sticker price is only good for commercial users since it's still cheaper than the dedicated industrial embedded boards, which I think is the new market the RPI company caters to. I haven't seen any embedded product company that doesn't incorporate RPis in its products now, or at least in their lab/dev/testing stage, so if you can sell your entire production stock to industrial users why bother making less money selling to consumers?

      • Mashimo 28 minutes ago
        I still use Pis in my 3d printers. Laptop would be too big, and a ESP could not run the software. "China clone" might work, but the nice part of the pi is the images available. It just works™

        I'm also currently building a small device with 5" touchscreen that can control a midi fx padle of mine. It's just so easy to find images, code and documentation on how to use the GPIO pins.

        Might be niche, but that is just what the Pi excels at. It's a board for tinkers and it works.

      • haunter 50 minutes ago
        Not just laptops but the used enterprise micro PCs from Dell, HP, and Lenovo. All the same small form factor with very low TDP You can have up to 32 or 64 GB RAMs depending on the CPU, dual or even triple disks if you want a NAS etc.
        • Mashimo 38 minutes ago
          I have seen quite a few the size of a Mac Studio / Intel Nuc, what are the device called that are the size of a pi?
        • joe_mamba 42 minutes ago
          yeah, depends on what the used market looks like where you live. Here I see way more laptops for sale for cheap than those enterprise thin clients.

          And the thin clients when they are for sale tend to have their SSDs ripped out by IT for data security, so then it's a hassle to go out and buy and extra SSD, compared to just buying a used laptop that already comes with display , keyboard, etc.

      • cyberax 25 minutes ago
        RPi will still have lower power consumption and is far more compact. And mechanically reliable.

        I'm in the market to replace my aging Intel NUCs, but RPi is still cheaper.

      • qsera 39 minutes ago
        >I can buy on the used market a ~2018 laptop with a 15W quad core CPU, 8GB RAM, 256 NVME and 1080p IPS display, that's orders of magnitude more capable..

        But it won't be as reliable, mostly motherboards won't last long.

        • joe_mamba 38 minutes ago
          Don't know what your source is for that, but that's not my experience, and i've had dozens of laptops through my hands due to my hobby.

          The ticking timebomb lemons with reliability or design issues, will just die in the first 2-4 years like clockwork, but if they've already survived 6+ years without any faults, they'll most likely be reliable from then on as well.

    • TazeTSchnitzel 55 minutes ago
      The Raspberry Pi probably still has the advantage of an actually robust firmware/software ecosystem? The problem with SBCs has always been that the software situation is awful. That was the Raspberry Pi's real innovation: Raspbian and a commitment to openness.
      • vachina 29 minutes ago
        > The problem with SBCs has always been that the software situation is awful

        Awful how? A SBC can take advantage of many software written from the dawn of x86.

        • sam_lowry_ 3 minutes ago
          x86 SBCs are niche, and ARM or RISC SBCs are too complex to figure out the boot process and the drivers for an average tinkerer.
    • ulnarkressty 1 hour ago
      That niche was long taken over by cheap Chinese SBCs, so they have to innovate somehow. Their only advantage that remains is the community.
    • crimsoneer 50 minutes ago
      Not everything needs to be for everyone. I think this is super cool - I run a local transcription tool on my laptop, and the idea of miniaturising it is super cool.
      • noodletheworld 38 minutes ago
        It is neat, and at 32GB it might be useful.

        Almost nothing useful runs in 8.

        This is the problem with this gen of “external AI boards” floating around. 8, 16, even 24 is not really enough to run much useful, and even then (ie. offloading to disk) they're so impractically slow.

        Forget running a serious foundation model, or any kind of realtime thing.

        The blunt reality is fast high memory GPU systems you actually need to self host are really really expensive.

        These devices are more optics and dreams (“itd be great if…”) than practical hacker toys.

  • Barathkanna 5 minutes ago
    As an edge computing enthusiast, this feels like a meaningful leap for the Raspberry Pi ecosystem. Having a low-power inference accelerator baked into the platform opens up a lot of practical local AI use cases without dragging in the cloud. It’s still early, but this is the right direction for real edge workloads.
  • rballpug 1 minute ago
    Catalog TG 211, 1000 Hz.
  • dwedge 1 hour ago
    > In practice, it's not as amazing as it sounds.

    8GB RAM for AI on a Pi sounds underwhelming even from the headline

  • speedgoose 47 minutes ago
    Is there any usefulness with the small large language models, outside perhaps embeddings and learning?

    I fail to see the use-case on a Pi. For learning you can have access to much better hardware for cheaper. Perhaps you can use it as a slow and expensive embedding machine, but why?

  • agent013 1 hour ago
    A good illustration of how “can run LLM” ≠ “makes sense to run LLM”. A prime example of how numbers in specs don’t translate into real UX.
  • teekert 4 minutes ago
    At this moment my two questions for these things are:

    1. Can I run a local LLM that allows me to control Home Assistant with natural language? Some basic stuff like timers, to do/shopping lists etc would be nice etc.

    2. Can I run object/person detection on local video streams?

    I want some AI stuff, but I want it local.

    Looks like the answer for this one is: Meh. It can do point 2, but it's not the best option.

  • cmpxchg8b 38 minutes ago
    8GB? What is this, an LLM for ants?
  • phito 1 hour ago
    Sounds like some PM just wanted to shove AI marketing where it doesn't make sense.
  • huntercaron 58 minutes ago
    Glad Jeff was critical here they need a bit of a wake up call it seems.
  • renewiltord 1 hour ago
    What’s the current state of the art in low power wake word and speech to text? Has anyone written a blog post on this?

    I was able to run a speech to text on my old Pixel 4 but it’s a bit flaky (the background process loses the audio device occasionally). I just want to take some wake word and then send everything to remote LLM and then get back text that I do TTS on.

    • geerlingguy 58 minutes ago
      Maybe not SOTA but the HA Voice Preview Edition [1] in tandem with a Pi 5 or some similar low-power host for the Piper / Whisper pipeline is pretty good. I don't use it but was able to get an Alexa/Google Home-like experience going with minimal effort.

      I was only using it for local Home Assistant tasks, didn't try anything further like retrieving sports scores, managing TODO lists, or anything like that.

      [1] https://www.home-assistant.io/voice-pe/

    • monocasa 57 minutes ago
      Wake word can be tiny. Like 10k weights and can run on an esp32 or similar with plenty of compute to spare.

      TinyML is a book that goes through the process of building a wake word model for such constrained environments.

  • kotaKat 20 minutes ago
    "For example, the Hailo 10H is advertised as being used for a Fujitsu demo of automatic shrink detection for a self-checkout."

    ... why though? CV in software is good enough for this application and we've already been doing it forever (see also: Everseen). Now we're just wasting silicon.

  • moffkalast 1 hour ago
    > The Pi's built-in CPU trounces the Hailo 10H.

    Case closed. And that's extremely slow to begin with, the Pi 5 only gets what, a 32 bit bus? Laughable performance for a purpose built ASIC that costs more than the Pi itself.

    > In my testing, Hailo's hailo-rpi5-examples were not yet updated for this new HAT, and even if I specified the Hailo 10H manually, model files would not load

    Laughable levels of support too.

    As another datapoint, I've recently managed to get the 8L working natively on Ubuntu 24 with ROS, but only after significant shenanigans involving recompiling the kernel module and building their library for python 3.12 that Hailo for some reason does not provide outside 3.11. They only support the Pi OS (like anyone would use that in prod) and even that is very spotty. Like, why would you not target the most popular robotics distro for an AI accelerator? Who else is gonna buy these things exactly?

  • simjnd 1 hour ago
    Ah yes, more AI e-waste