Tag Archives: LLM

LLM – Optical problem solving now works

Finally got the PDF OCR flow working, and the LLM can now solve math problems. Only grey diagrams are still hard for it to recognize properly.

I used a math training exam that my son is currently working on in preparation for Gymnasium.

And even the 27B model is sufficient to solve this level of math.

Yet another Mac Studio M3 Ultra this time with 512GB

Did I say enough…..
meh….

Okay lesson learned, never enough.
My “1 month old” M3 Ultra 256GB went out of memory running all those models and podman containers in parallel.



My current setup is:
OpenWebUI:
-> LM Studio, Aya-Vision-32b,
-> ComfyUI workflow with t5xxl, Flux1.dev, llama-3.1
-> MCP proxified: Searxng, wikipedia, docling, context7, time, memory, weather, sequential-thinking
Podman: 24 containers including supabase, wikijs, watchtower…

Also I discovered that I can use OpenWebUI, SwarmUI, exo and even mlx
to distribute workload across both Mac Studios via 80GBs thunderbolt 5 bridging.

And with the orange clown, you never know if there will be a new M4 Ultra next year at all.

LLM full throttle on M3 Ultra 84 Watts !

What a crazy efficiency monster.
Running a LLM on all 80 GPU cores and the system is only drawing 84 Watts…..

Nvidia must be crying at night.

and this is only M3, M4 is even more efficient, but yet not available as an ultra fusion variant, and end of year Apple will manufacture in 2nm



Upgraded my old M1 mini to Mac Studio M3 Ultra 256GB

Finally the latest Mac Studio was released, unfortunately only with M3 chip instead of a M4, but I simply waited too long already.
Let’s see how this one compares to the previous M4 Pro I bought for my son in regards to LLM’s.

This one will be used with Podman Desktop, LM Studio and hopefully be fast enough to also handle voice recognition and rendering in realtime.

256GB VRAM will be more than enough for my use cases, 96GB was simply not enough as I already saw that on the 64GB M4.