Thanks Dennis :] --------------------------------------- Jeremy Schwartz Starmark Interactive Developer [EMAIL PROTECTED] 954-761-1600
> From: "Dennis Stout" <[EMAIL PROTECTED]> > Date: Tue, 4 Feb 2003 00:14:30 -0900 > To: "Jeremy Schwartz" <[EMAIL PROTECTED]>, <[EMAIL PROTECTED]>, "Andy > Lester" <[EMAIL PROTECTED]> > Cc: <[EMAIL PROTECTED]>, <[EMAIL PROTECTED]> > Subject: Re: Search for a string > > Sheesh. > > Wil ya'll just help a man with a perl problem instead of battering him with > other ways to do it? > > Sometimes people like ot pose a challenge to themselves and see if it can be > done. > > Instead of being counterproductive and refering peopel to other things, help > the man! > > I wish I could but at the moment, I'm trying to get some sleep.. > > Good Night, > Dennis Stout > > >> I am not handling this at the Apache level because I have about 40+ >> different sites running on this one box. I want to be able to run server >> wide statistics from the 'master' log file, then I want to strip it apart >> into the separate pieces, pipe that piece to Analog and dump the results >> into the vhost folder for the site owner to view. >> >> I don't want to have to collect and run through 40 individual files, I want >> to have the system do the work creating temporary vhost log files and >> processing them. >> >> That mod to my shell script is kind of what I am looking for. I guess I >> wasn't sure if I could do that with Perl as well. >> --------------------------------------- >> Jeremy Schwartz Starmark >> Interactive Developer >> [EMAIL PROTECTED] 954-761-1600 >> >>> From: [EMAIL PROTECTED] >>> Date: Mon, 3 Feb 2003 12:32:05 -0600 >>> To: "Jeremy Schwartz" <[EMAIL PROTECTED]>, "Andy Lester" > <[EMAIL PROTECTED]> >>> Cc: [EMAIL PROTECTED], [EMAIL PROTECTED] >>> Subject: Re: Search for a string >>> >>> >>> ------------------------------------------------ >>> On Mon, 03 Feb 2003 13:09:47 -0500, Jeremy Schwartz <[EMAIL PROTECTED]> >>> wrote: >>> >>>> Not trying to reinvent the wheel. >>>> >>>> I am using Analog for the analysis. >>>> >>>> I am trying to split the server combined log into individual vhost logs. > I >>>> can then run each through Analog to produce individual reports. >>> >>>>> Don't reinvent the wheel. There are a number of fine log analysis >>>>> utilities, such as analog. >>>>> >>>>> xoa >>> >>> Out of curiousity is there a reason why you are not handling this at the >>> Apache level? Each vhost can have its own set of logs at the start that > then >>> would not need to be pulled apart. Is this a possible scenario for you > going >>> forward? (granted it doesn't help now). It would seem that your task > would be >>> better handled with shell script possibly since you already have the > command >>> line for creating the file(s) from the main log, so then just wrap that >>> command in a foreach that takes your directory names as input. >>> >>> Something along the lines of: >>> >>> #!/bin/sh >>> >>> for dir in `ls -1 /webroot/`; do >>> cat /var/log/httpd/access_log | grep "$dir" > >>> /var/log/httpd/access_log_$dir >>> done >>> >>> I am no shell hacker and the above is untested, but you get the idea. In >>> general Perl would not be a good choice for performing something so simple >>> that already has a command line solution available. >>> >>> If you were going to do it in Perl, rather than looking for each vhost in > the >>> log file, you would be better off unpacking or splitting, etc. the log > line >>> and storing that line to an array that is associated with the particular > vhost >>> in the log line and then printing each vhost's array to a file, or you > would >>> have to open a filehandle for each vhost at the beginning of the script > and >>> then just print the line to whichever filehandle is associated with a >>> particular vhost. Stepping through every line of the log file foreach of > the >>> vhosts in Perl would probably be a really bad way to handle things. >>> >>> I would still suggest letting Apache do the splitting by not storing one > main >>> log with all vhost content, it is much easier to put the logs back > together to >>> get a complete picture than it is to disect them after the fact. >>> >>> http://danconia.org >>> >>> -- >>> To unsubscribe, e-mail: [EMAIL PROTECTED] >>> For additional commands, e-mail: [EMAIL PROTECTED] >>> >> >> >> -- >> To unsubscribe, e-mail: [EMAIL PROTECTED] >> For additional commands, e-mail: [EMAIL PROTECTED] >> > > > -- > To unsubscribe, e-mail: [EMAIL PROTECTED] > For additional commands, e-mail: [EMAIL PROTECTED] >