Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

High overhead of power measurements for CPU inference measurements #236

Open
psyhtest opened this issue May 5, 2021 · 11 comments
Open

High overhead of power measurements for CPU inference measurements #236

psyhtest opened this issue May 5, 2021 · 11 comments
Assignees
Labels
Investigate Investigate/Research
Milestone

Comments

@psyhtest
Copy link
Contributor

psyhtest commented May 5, 2021

During the MLPerf Inference v1.0 round, I noticed that the power workflow when used with CPU inference occasionally seemed to incur a rather high overhead (~10%), for example:

Here, ArmNN is faster than TFLite but takes a big hit under the power workflow. TFLite, however, is not affected.

@psyhtest psyhtest self-assigned this May 5, 2021
@psyhtest psyhtest added the Investigate Investigate/Research label May 5, 2021
@psyhtest psyhtest added this to the v1.1 milestone May 5, 2021
@psyhtest
Copy link
Contributor Author

psyhtest commented May 5, 2021

The observed behaviour is consistent across several runs:

@s-idgunji
Copy link
Contributor

Hi Anton - On Jetson AGX, we can check our runs with and without power to see if we see an impact of the power related SW overheads in communicating to the director PC. I will comment back on this issue.

@psyhtest
Copy link
Contributor Author

@s-idgunji When running inference on the GPU, the CPU is typically not fully utilised, so you might not notice any difference. When running inference on the CPU in full-blast, the effect may be more pronounced.

@s-idgunji
Copy link
Contributor

@psyhtest - Thanks, we were not able to notice a difference when we had workload going on GPU on the Xavier config that we submitted, but I agree that when you are using CPU only, then the effect may show up. We can independently try and replicate the behavior you observed. But we'll need to close on this soon and propose a fix that can go in so that we can validate prior the freeze date.

@TheKanter
Copy link
Contributor

Anton - RP4 is pretty low-end CPU. I don't see any way around some sort of overhead. Is the overhead consistent or are you saying it's unpredictable?

@psyhtest
Copy link
Contributor Author

psyhtest commented Jun 2, 2021

The overhead seems to be consistent. What's worrying is that it shows up with ArmNN on Xavier which is like 8x more powerful than RPi4.

@s-idgunji
Copy link
Contributor

Anton - Can you try on another system , perhaps Intel Celeron or Core based. Also , what can we suggest as a fix ? If we want to repro on our system at NVIDIA what specific config , benchmark/scenario would likely give the highest difference. We could test it out to see if it repros.

@araghun
Copy link
Contributor

araghun commented Jun 16, 2021

Power WG Thoights:

For v1.1: Try and solve it opportunistically. It is not a gating items.
For further revisions: We could take a look into it.

This is mainly due to the resources available to solve this problem. Main issue is there is no minimum bar for the software flow.

@araghun
Copy link
Contributor

araghun commented Jul 14, 2021

MLPerf Power WG: This will be punted to after v1.1. issue will remain open.

@s-idgunji
Copy link
Contributor

@psyhtest - perhaps you want to close this if the issue is not valid anymore, or we need to revisit in the WG. Quite an old issue.

@arjunsuresh
Copy link
Contributor

Is this issue still happening?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Investigate Investigate/Research
Projects
None yet
Development

No branches or pull requests

5 participants