On 12/4/13, Matt Hope <matt.h...@gmail.com> wrote:
> Random tips:

cp -t argument sometimes useful

rsync -avu src/ host://dest/ might be preferable

when find's -print0 also use:
 [e]grep's -z and -Z (or -zZ) options
 xarg's -0 option

Something I cut and pasted off the internet years ago:
  Way to solve it if you can't use GNU utilities:
    find -name "*.txt" | sed 's/"/\\"/g;s/.*/"&"/' | xargs grep "whatever"
  This escapes out all the double quotes in the pathname and
  then wraps the entire pathname in double quotes.

And if your directory is excessively large (this was a reminder to
myself years ago, probably no longer applicable with the modern
filesystems and proper direntry hashing):
For very large directories (12K files), for some reason piping find
into sed, can cause each iteration (ie. each file) to take a second or
two each. Bizzare. So use the following:
ls > zzz; then do the following; cat zzz |xargs sed -i -e
"s%http://wiki.jboss.org/wiki/Wiki.jsp?page=%%g;"; -e "s%<base
href=\"http://www.jboss.com/\"/>%%g;"

If you end up on your char fishing expedition, you might prefer to use
the "ls > tmp.file" form, followed by some sed or perl.

Good luck.
-- 
SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/
Subscription info and FAQs: http://slug.org.au/faq/mailinglists.html

Reply via email to