Magic Banana does it again ... with "tr -s ' ' < file | sed 's/ $//' | sort
-u"
I plugged in the filename thusly:
> tr -s ' ' < HolkitsorN.txt | sed 's/ $//' | sort -u > HolkitsorN-MBd.txt
With the result that the original 13.1 kB shrank to 8.8 kB, without any
residual duplicate lines.
At Magic Banana's suggestion, I applied sort -u to the test file in my
original
posting, with a perfect elimination of the extra files, but when I tried the
same approach on another set of data that I had first sorted with LibreOffice
Calc, sort -u reduced the numer of rows from 67 to 46, but
> jaholper1.example.com 95.182.79.24
> jaholper1.example.com 95.182.79.24
> jaholper1.example.com 95.182.79.33
> jaholper1.example.com 95.182.79.33
> jaholper4.example.com 109.248.200.4
> jaholper7.example.com 109.248.203.131
> jaholper7.example.com
As apache server software on shared servers routinely performs hostname
lookups
on data requests made to the hosted domains on their servers, I'm compiling a
database of the thousands of example.com hostnames that are on the Internet.
I've reached an impasse: LibreOffice's Calc spreadsheet