Thanks y_ich,
I had no idea how much more powerful the neural engine was compared to the gpu. I am glad your app either effectively uses both, or that it’s not a big deal if it isn’t using the gpu. I’m a little surprised the development tools don’t let you dig in and find out more.
Regardless I am delighted with 200 visits per second on the 40 block katago engine. So thank you again!
And I’ll repeat the caveat for others, don’t upgrade to the Pro version of iPhone 13 for this app, the regular is good enough. (The zoom camera was enough reason for me.)
y_ich wrote:wineandgolover san,
I have no way to know inside A15 and Core ML, but your experiment is quite reasonable.
A15's Neural Engine runs at 15.8FLOPS (though Apple says TOPS instead of FLOPS).
A15's GPU with 5 cores seems run at about 2.5FLOPS with FP16.
It means that 1 core runs at about 0.5FLOPS.
If Core ML uses both Neural Engine and GPU effectively, A15 with 5 cores has total 18.3FLOPS, while A15 with 4 cores has 17.8FLOPS.
The difference is 0.5FLOPS as described above, 0.5/18.3 = 0.027 ~ 3%.
I think that the match with your test is by chance.
Thermal variation may larger than it.
FYI