I have directory with cca 26 000 files and I need to grep in all these files. Problem is, that I need it as fast as possible, so it's not ideal to make script where grep will take name of one file from find command and write matches to file. Before "arguments list too long" issue it took cca 2 minutes to grep in all this files.
Any ideas how to do it?
edit: there is a script that is making new files all the time, so it's not possible to put all files to different dirs.
Grep Performance – Search in Thousands of Files Efficiently
filesgrepperformance
Best Answer
With
find
:(
-type f
is to only search in regular files (also excluding symlinks even if they point to regular files). If you want to search in any type of file except directories (but beware there are some types of files like fifos or /dev/zero that you generally don't want to read), replace-type f
with the GNU-specific! -xtype d
(-xtype d
matches for files of type directory after symlink resolution)).With GNU
grep
:(but beware that unless you have a recent version of GNU grep, that will follow symlinks when descending into directories). Non-regular files won't be searched unless you add a
-D read
option. Recent versions of GNUgrep
will still not search inside symlinks though.Very old versions of GNU
find
did not support the standard{} +
syntax, but there you could use the non-standard:Performances are likely to be I/O bound. That is the time to do the search would be the time needed to read all that data from storage.
If the data is on a redundant disk array, reading several files at a time might improve performance (and could degrade them otherwise). If the performances are not I/O bound (because for instance all the data is in cache), and you have multiple CPUs, concurrent
greps
might help as well. You can do that with GNUxargs
's-P
option.For instance, if the data is on a RAID1 array with 3 drives, or if the data is in cache and you have 3 CPUs whose time to spare:
(here using
-n1000
to spawn a newgrep
every 1000 files, up to 3 running in parallel at a time).However note that if the output of
grep
is redirected, you'll end up with badly interleaved output from the 3grep
processes, in which case you may want to run it as:(on a recent GNU or FreeBSD system) or use the
--line-buffered
option of GNUgrep
.If
pattern
is a fixed string, adding the-F
option could improve matters.If it's not multi-byte character data, or if for the matching of that pattern, it doesn't matter whether the data is multi-byte character or not, then:
could improve performance significantly.
If you end up doing such searches often, then you may want to index your data using one of the many search engines out there.