site banner

Small-Scale Question Sunday for August 10, 2025

Do you have a dumb question that you're kind of embarrassed to ask in the main thread? Is there something you're just not sure about?

This is your opportunity to ask questions. No question too simple or too silly.

Culture war topics are accepted, and proposals for a better intro post are appreciated.

2
Jump in the discussion.

No email address required.

Has anyone here used the new AMD Ryzen AI Max+ 395 all-in-one chipset for running local AI? Like Apple M, it uses a unified memory model, so you can dedicate up to 96GB to GPGPU/NPU tasks.

It's a laptop cpu...? Do people buy expensive laptops in order to run local llms on them? Just curious.

You see it on /r/locallama a bit. It’s usually slow, but for async tasks that may not matter as much, and being able to run higher bpw helps a lot.

If you can use online services, they’ll absolutely paste most local llms at this scale, but there’s a lot of use cases where online services aren’t an option, or philosophically unpalatable.