Now i can run llava-1.5-7b model on Macbook pro M2 Max without any convert!
He actually understood that i'm a fencer 🤺
MLX 😍
Thanks for all great works!
Awni Hannun
Prince Canuma
Apple MLX: Phi-3 running locally on a VisionPro with VisionOS 1.2 Beta 3!
Fully offline, pretty fast! 22.25 t/s 🔥
Credit to Awni Hannun for the special quantized version for MLX 😉
In the code I used displayEveryNTokens = 3 to make streaming more 'continuous'.