I need to maximize filesystem sequential read performance and I am willing to do unreasonable things as long as I don't have to spend money.
Context: I have a bunch of data files (1-2GB each) that I need to read sequentially and process as I read them. The files need to be processed in multiple passes in different orders, so basically I am sequentially reading a bunch of big files over and over.
Files are stored on a two-disk (3Gb SATA) software RAID-0 (which strangely doesn't seem to help much). I also only have 1GB of memory, significant part of which is used by data processing code.
Could somebody suggest ways to optimize my RAID and filesystem configuration that would help with sequential reads? I am willing to trade-off pretty much everything for read performance.
So far XFS with default options seems to perform best overall. I tried to configure my RAID with chunk sizes varying from 4 to 512kB and it didn't seem to make that much difference either.
Could somebody help?




