Thanks Wolf.
It is working now.
cheers,
Purvee
On Mon, Jul 21, 2014 at 11:07 AM, WFB wbi...@gmx.at wrote:
Hi Purvee,
perl -e 'xyz' works fine on Unix, however on Windows you must use
instead.
perl -e print qq(Hello \n)
greetings,
wolf
On 21 July 2014 07:24, Purvee Vora
I have self-learned Perl about a couple of years ago and I am also having a
similar use of Perl, so I often forget how I did things. I must also say
that I am about to be 50 years old next December and I do not have any
technical education as my degree is in psychology. I use
On Mon, 21 Jul 2014 15:01:10 +0200
mimosin...@gmail.com wrote:
I have self-learned Perl about a couple of years ago and I am also
having a similar use of Perl, so I often forget how I did things. I
must also say that I am about to be 50 years old next December and I
do not have any technical
Greetings,
I self learned perl four years ago in the year 2010. The primary book from
where I learned Perl programming is Learning Perl
http://www.amazon.com/Learning-Perl-Randal-L-Schwartz/dp/1449303587/ref=sr_1_1
Then, I trained about 200+ students on Perl programming in the last 4 years
I am now 73 years old and retired one year ago. I used Perl extensively from
1998 until I retired. It can be hard to remember all the tricks, even when you
use or write Perl often. When I learned or developed a useful technique, I
would keep the program or a bit of it around. I have lots of
I strongly recommend to use a cheat sheet or pocket reference books as this
one http://shop.oreilly.com/product/0636920018476.do. Personally, I use a
french language Perl Moderne pocket reference book (http://perlmoderne.fr/)
that helps me almost daily.
I really love this way to learn because
I checked CPAN for remove duplicate lines and only found Code::CutNPaste
which doesn't sound like what I want. So I will build what I want although
I'm sure it's out there somewhere.
I have several files and I attached one of them. I want to sort the file
and remove duplicate lines. The file is a
Sheman,
Do it the perl way, hash it.
1. Read both files
2. Put the lines into the key of a hash.
* While (reading files)
* $hash{$_}++;
Now you can print/process your hash key. There’s shouldn’t be duplicates.
From: Sherman Willden
On Mon, 21 Jul 2014 16:01:05 -0600
Sherman Willden sherman.will...@gmail.com wrote:
I have several files and I attached one of them. I want to sort the
file and remove duplicate lines.
If you're running bash:
sort -u file output_file
--
Don't stop where the ink does.
Shawn
--
To
On Mon, Jul 21, 2014 at 10:05:29PM +, Danny Wong (dannwong) wrote:
Do it the perl way, hash it.
Or do it the unix way:
$ sort -u filename
The -u means unique.
You also have some lines that differ only in case, so you might prefer:
$ sort -uf filename
The -f means fold, which ignores
Thank you all. I got off of windows 7 and went to my Ubuntu. Great stuff.
Now I have to go back to windows 7 to take the course.
Sherman
On Mon, Jul 21, 2014 at 4:14 PM, Paul Johnson p...@pjcj.net wrote:
On Mon, Jul 21, 2014 at 10:05:29PM +, Danny Wong (dannwong) wrote:
Do it the perl
11 matches
Mail list logo