-
Notifications
You must be signed in to change notification settings - Fork 25
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Enhancement: write a TD;DR file on how to run tests #5
Comments
Hello smclinden, To start, we would recommend that you convert those ASCII "0" and "1" bytes into single binary bits, and strip out all whitespace (spaces, tabs, newlines, etc.) so that you are processing ONLY a raw binary file. That will require less memory to process as well. |
That's easy enough. But what would be the arguments to sts? |
STS would choke on that much data. The spec says 1,000,000 bits. You file
is 100s of times bigger.
…On Mon, Feb 8, 2021, 7:08 AM smclinden ***@***.***> wrote:
So I have a file of 108196287 lines of 8-byte strings converted to ASCII
0/1.
How do I run the tests on this file?
I was never able to get sts-2.1.2 to run without alloc errors.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
<#5>, or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABYMCGPTOFCSL2C64T4R4MDS57465ANCNFSM4XJFDHKA>
.
|
I could shorten it. The issue is that the data is a set of 8-byte codes that are supposed to be random but I have reason to believe that the PRNG is flawed (it appears that someone has figured out how to generate valid additional codes based upon what is already known). I can break the data up but I would need to know the appropriate strategy for doing so. |
I have a similar file with 1000 datapoints of 32-bits each. Should I create a newline with 32-bits value for each data point? What should be the value of datastream in './assess '? |
The problem is that you only have 32000 bits if data, which is a very small sample on which to make a meaningful measurement. If I were trying to look at the quality of the data, I would try for at least 1 000 000 such 32-bit data points. However if you insist on testing such a small amount of data: # generate 32000 bits from a lower quality source /dev/random
# bs=4 is 4 bytes or 32 bits
dd if=/dev/random of=binary.file bs=4 count=1000
/usr/local/bin/sts -S 1000 -i 32 binary.file assuming that You will notice that a number of sub-tests are disabled. For example the above run, the following warnings were produced:
So about 1/2 of the models cannot even get started to evaluate data due to the small data sample size. If you had 1 000 000 32-bit data points as in: ```sh
# generate 32000000 bits from a lower quality source /dev/random
# bs=4 is 4 bytes or 32 bits
dd if=/dev/random of=binary.file bs=4 count=1000000
/usr/local/bin/sts -S 1000000 -i 32 binary.file then the We hope this helps @nivi1501 |
We do plan to write a |
So I have a file of 108196287 lines of 8-byte strings converted to ASCII 0/1.
How do I run the tests on this file?
I was never able to get sts-2.1.2 to run without alloc errors.
The text was updated successfully, but these errors were encountered: