Performance of llama.cpp on Apple Silicon A-series #4508
Replies: 19 comments 32 replies
-
iPhone 15 Pro (A17 Pro) ✅
|
Beta Was this translation helpful? Give feedback.
-
iPhone 15 Pro Max (A17 Pro) ✅
Tested under iOS 17.3 Developer beta 1 (21D5026f) |
Beta Was this translation helpful? Give feedback.
-
iPhone 12 mini (A14) ✅tinyllama:
phi-2:
Tested under iOS 17.1.2 (21B101) |
Beta Was this translation helpful? Give feedback.
-
Some additional info with memory and relevant devices.
|
Beta Was this translation helpful? Give feedback.
-
iPhone 13 Pro (A15) ✅
also a model |
Beta Was this translation helpful? Give feedback.
-
iPhone 14 Pro (A16) ✅
|
Beta Was this translation helpful? Give feedback.
-
iPhone 12 (A14) 🟨
|
Beta Was this translation helpful? Give feedback.
-
can anyone tell me what does the output metric (t/s) mean? tokens per second or what? |
Beta Was this translation helpful? Give feedback.
-
Can anyone tell me about llama 1b download link? I can't find it on HF or not sure which is. |
Beta Was this translation helpful? Give feedback.
-
Hi, I was trying to load starcoderbase-3b-GGUF. It is not getting loaded in iphone 15 pro simulator. It is stuck with |
Beta Was this translation helpful? Give feedback.
-
iPhone SE (3rd Generation), A15 2+4 CPU, 4 GPU, 4 GB of RAM
|
Beta Was this translation helpful? Give feedback.
-
What data/prompts are used for this? |
Beta Was this translation helpful? Give feedback.
-
I have run llamma.cpp on ios device (iphone) described here. But models are giving garbage response. what am I doing wrong? |
Beta Was this translation helpful? Give feedback.
-
Would it be possible to update these instructions for a recent version of XCode? I get a simple error that I can't quite figure out: |
Beta Was this translation helpful? Give feedback.
-
Same error as kinchahoy. Getting the same error. |
Beta Was this translation helpful? Give feedback.
-
Hi, I was trying to deploy llama.swiftUI on IOS 15. |
Beta Was this translation helpful? Give feedback.
-
why is it that we need to use the "Release" build? Thanks |
Beta Was this translation helpful? Give feedback.
-
Hi, I'm new to the llama cpp. The llama.swiftui works well in my local machine by using the default downloadable models, but it failed to start after I copied the quatanized meta llama 3.2 1B, which is created by using HuggingFace https://huggingface.co/spaces/ggml-org/gguf-my-repo, to ![]() It showed errors from console:
Would you please point out what went wrong? |
Beta Was this translation helpful? Give feedback.
-
iPhone 16 pro, A18 Pro 2+4CPU, 6GPU, 8 GB of RAMThe first thing to do after I got my new phone is to run llama benchmark 👍
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Summary
🟥 - benchmark data missing
🟨 - benchmark data partial
✅ - benchmark data available
TinyLlama 1.1B
Cores
Cores
[t/s]
[t/s]
[t/s]
[t/s]
[t/s]
[t/s]
Phi-2 2.7B
Cores
Cores
[t/s]
[t/s]
[t/s]
[t/s]
Mistral 7B
Cores
Cores
[t/s]
[t/s]
Description
This is a collection of short
llama.cpp
benchmarks on various Apple Silicon hardware. It can be useful to compare the performance thatllama.cpp
achieves across the A-Series chips. Similar collection for the M-series is available here: #4167Instructions
iPhone 13 mini ✅
Footnotes
https://en.wikipedia.org/wiki/Apple_A14 ↩ ↩2 ↩3
https://en.wikipedia.org/wiki/Apple_A15 ↩ ↩2 ↩3 ↩4 ↩5 ↩6 ↩7 ↩8 ↩9
https://en.wikipedia.org/wiki/Apple_A16 ↩ ↩2 ↩3
https://en.wikipedia.org/wiki/Apple_A17 ↩ ↩2 ↩3
Beta Was this translation helpful? Give feedback.
All reactions