-
Notifications
You must be signed in to change notification settings - Fork 467
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[SYSTEMDS-???] SchemaApply Performance Tests #1869
Conversation
Improve Column combining when the different statistics have equal cost (for instance if it is constants.)
after:
|
And a shortcut optimization:
|
Initial in memory performance test of the Compression and processing speed,
FYI @mboehm7 , Do you think this is the right direction? Following tests will confirm if this also works when writing to disk. |
On So010, with a 1000 x 1000 matrix and 48 threads in use: we get 40+-8 GB input processed with MatrixVector multiplication with a peak MemoryBandwith of 200 GB
|
965aaa7
to
abebe83
Compare
This commit extends the performance Jar for measuring internal functions. In specific this commit adds functionality to measure bandwidth utilization of individual operations.
abebe83
to
1bddf49
Compare
This PR contains code to compare schema compression with our standard compression.
There are good indications that the schema apply is much faster, in the currently supported schemas.
In the following tables the update scheme takes the compressed scheme from one of the previously compressed blocks, and update the scheme to enable compression of new given blocks.
The apply scheme takes the scheme and applies to incoming uncompressed MatrixBlocks.
Update and Apply does both.
From Empty, takes an empty DDC single column scheme (one group per column) and materialize a new scheme for each block given and then applies it to the given MatrixBlock returning a compressed block.
Scaling number of unique values;
Scaling number of columns