Hi I'm running the dieharder test suite (this) to test an algorithm, but I've found somethings that I don't understand.
This is the command that I run to test my binary file
dieharder -f test -a -g 201
I've tested my algorithm on file of different size: 30Mb, 300Mb, 3Gb and the result are that I fail a lot of test on the 30Mb file, less in the 300 Mb (still a lot), and in the 3Gb I pass all the tests.
Isn't it strange? Shouldn't the tests produce wrost results if the file is larger? (or the weakness become less visible) If it's normal, can this be considered a weakness in the algorithm?
Also shouln't the tests consider the file length? and so "adapt" to it
EDIT: If statistical anomalies also go down if a larger test file is used what is the optimal size that I should use to do the tests?
PS when I talk about weakness I mean that it's weak statistically speaking