-
Notifications
You must be signed in to change notification settings - Fork 941
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to reproduce iOS benchmark values #243
Comments
Hello! Your setup sounds accurate to me. Which version of Stable Diffusion are you benchmarking? |
I made no changes to the example app, and it seems to use |
That is also correct. A few things:
|
Hmm, that sounds even faster than what we published (2.72 vs 2.3 iter/sec) and it should have finished in ~8 seconds with that throughput. I will wait for Pedro to repro his measurements from June and also rerun our measurements on Seed 5 this week. |
@kanaukou-google Oh one more thing, could you please verify that |
I no longer have access to the iPhone 14 Pro I used for the tests, but I repeated them on my iPhone 13 Pro running iOS 17 beta 7 (21A5319a). Some observations:
To reduce ambiguity, I pushed this branch to replicate the benchmark conditions (to the best of my recollection) using the latest code. Using that branch, I got the following results on 5 consecutive runs using Xcode 15.0 beta 7 on iPhone 13 Pro running iOS 17 beta 7 (21A5319a):
Also observe that I'm running tests after reboot, waiting for the device to cool, and detached from Xcode. |
I can also repeat the tests on seed 5 of Xcode if that's useful. |
I checked out @pcuenca's PR and tried reproducing it on a cool rebooted 14 Pro beta 4 detached from Xcode. Below are results for 5 consecutive runs (meaning 5 times pressed 'Generate' button after processing is complete without restarting the app or changing the prompt).
The results look even better than the benchmark values! I wonder if such a consistency of .1s for time and .01s for it/s values for 5 runs is expected, though? |
Thank you both for the time spent on this @pcuenca @kanaukou-google! Our inference stack is consistently improving! We will rerun our internal benchmarks with the latest public seed of iOS 17 and update our numbers. |
No problem, glad we figured this out! Just one quick question, what would be the best approach to approach the benchmarking in future? Looks like some changes from @pcuenca's PR need to be applied to get proper results. |
That's a good point @kanaukou-google! I'll add a |
I am unable to reproduce claimed benchmark values for running Stable Diffusion model on iOS devices.
Configuration
Steps to reproduce:
Could anyone kindly clarify is there anything specific I need to adjust in my setup in order to reproduce the results?
The text was updated successfully, but these errors were encountered: