New pages

Jump to navigation Jump to search
New pages
Hide registered users | Hide bots | Hide redirects
  • 14:23, 13 April 20242024 llama.cpp 7840u (hist | edit) ‎[5,364 bytes]DavidM (talk | contribs) (Created page with "I briefly had a Macbook M3 Max with 64GB. It was pretty good at running local LLMs, but couldn't stand the ergonomics and not being able to run Linux, so returned it. I picked up a Thinkpad P16s with an AMD 7840 to give Linux hardware a chance to catch up with Apple silicon. It's an amazing computer for the price, and can run LLMs. Here's how I set up llama.cpp to use ROCm. Install ROCm, set an env variable for the 780m: <code>export HSA_OVERRIDE_GFX_VERSION=11.0.0</co...")