Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Yes, simply compiling the provided code and running it against provided data set does not really count as replication.

Writing your own code and creating your own dataset is the simplest way to rule out the situations where there is something fishy in either the original code or data. And the paper itself should contain enough details to make it possible to recreate the experiment this way.



I think compiling the provided code and running it against the provided data set does do something---you know the code reports on the data. If you get a different result with the provided code and data, then there's something different with your environment vs. the original researcher, perhaps the rounding mode, or some assumption [1]. Once that's straightened out, then you can use new data and see if you can replicate the result with the provided code and new data.

Think of the provided data as a sanity test.

[1] I recently fixed a bug wherein I was inadvertently relying upon Linux-specific behavior that failed when tested under Solaris.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: