Tag: AI
-
Back with a llama
TLDR; read this post if you want Llama 3 running locally on your ARM based Mac (aka M1, M2, M3) with inference speeds of 30+ tokens per second & GPU cores on full throttle. Less than 2 weeks ago, AI @ Meta released Llama 3. Since then, I have been frantically working on getting the…