Discussion about this post

User's avatar
Laurence Anthony's avatar

Fantastic work, Simon! Apple MLX really is amazing, isn't it? Any thoughts on Llamafile and how your LLM library compares to that?

Expand full comment
KimSia Sim's avatar

Thank you for this, Simon

I’m a Python/Django dev on Mac M1 Max (32 gb ram) and have been using Claude sonnet extensively at work via cursor

Any thoughts on how close this can get to replacing sonnet ?

I noted you wrote that llama 70b can run locally on a 64gb ram and get close to gpt 4o level

I’m thinking of perhaps some kind of divide n conquer

1. Consider upgrading to new Mac with more ram

2. Consider using your suggested setup but for easier queries

Your thoughts about these two paths

Expand full comment
3 more comments...

No posts