M4 MAX MacBook Pro BENCHMARKED: Deepseek v3 vs Qwen, Phi-4 and Llama on Ollama

Опубликовано: 06 Январь 2025
на канале: IndyDevDan
23,500
687

LET'S UNBOX THE M4 MAX MacBook Pro! ngl the m4 max is an absolute beast for LLMs. We BENCHMARKED it against Deepseek v3 with Ollama SLMs like Llama3.2, Qwen2.5, Falcon3-10b, and Phi-4.

Ready to see where the limits of local LLMs (SLMs) are on a top of the line M4 Max MacBook Pro (128GB RAM + 16 Cores + 4TB)?

👍 Resources:

Benchy Codebase: https://github.com/disler/benchy
Principled AI Coding: https://agenticengineer.com/principle...
Previous 2025 Predictions Video:    • AI Coding is here... What’s next? My ...  
Four Levels of the Prompt:    • Prompt Engineering Master Class for E...  
Promptfoo Testing Framework (Best Prompt Format):    • BEST Prompt Format: Markdown, XML, or...  
Deepseek: https://www.deepseek.com/

With Llama4 right around the corner it's time to get ahead of the curve and prepare our benchmarks. Benchmarking is critical for success in the Generative AI Age. Why? Because benchmarks tell you what you can do with local and cloud models. They literally tell you when a model can solve the problem you want solved. In this video we look at common patterns and mental models you can use to build powerful, useful, actionable benchmarks for your language models.

🚀 In this video, we put the M4 MAX MacBook Pro to the ultimate test against the new giant: Deepseek v3! We also compare the M4 MAX to the M2 Max to see if the M4 MacBook Pro is really worth the money (teaser: it is).

💻 Watch as we run powerful local models like Llama3.2, Llama4, Qwen, and Phi-4 on Ollama, and see how the M4 Max stacks up. Using tools like Promptfoo and our own Benchy codebase, we analyze tokens per second, accuracy, and overall performance.

🔬 We'll explore whether the M4 Max is truly the best machine for running local models and small language models (SLMs). Learn about the significance of LLM benchmarks and how they impact AI engineering and development.

🔥 Join me, IndyDevDan, as we uncover the capabilities of the M4 Max MacBook Pro and discuss what this means for AI enthusiasts, developers, and anyone interested in pushing the boundaries of tech in the Generative AI Age.

Stay focused and Keep building.

📖 Chapters
00:00 Unboxing M4 MAX MacBook Pro
01:48 Prepare for local LLMs with benchmarks
02:38 M4 Max vs M2 Max
05:45 Llama3.2:1b M4 vs M2
07:13 Falcon3:10b M4 vs M2
08:55 Qwen2.5-coder:32b M4 vs M2
10:50 Qwen2.5:72b M4 vs M2
13:53 M4 Benchmarking against Deepseek v3
30:25 SLM limitations

#macbookpro #aiagent #promptengineering