version of the mysql
database...
-Original Message-
From: [EMAIL PROTECTED] [SMTP:[EMAIL PROTECTED]]
Sent: Thursday, August 23, 2001 4:23 PM
To: [EMAIL PROTECTED]
Cc: [EMAIL PROTECTED]; [EMAIL PROTECTED]
Subject: Re: security: private or public pages
Briggs, Gary wrote
Because that way you can search through just the titles, etc, because of a
bitwise OR operation done on the weight.
Gary (-;
-Original Message-
From: Andre Pfeiler [SMTP:[EMAIL PROTECTED]]
Sent: Monday, August 20, 2001 10:13 AM
To: [EMAIL PROTECTED]; Andre Pfeiler
Subject:
Hi,
There's a limit to the amount of content it'll download and indx for each
individual page. IIRC, it's in the indexer.conf file, but it may be compiled
in...
Gary (-;
-Original Message-
From: [EMAIL PROTECTED] [SMTP:[EMAIL PROTECTED]]
Sent: Thursday, July 26, 2001 3:45 PM
To:
As threatened the other day, I've coded cookie support into the indexer. I
also implemented arbitrary header strings while I was at it...
Anyway, it reads columns called cookie_string and header_string out of
the server table, so you'll need to modify it with something like
MODIFY server ADD
Right, people
On a per-server basis, I wrote a patch before the weekend for cookie
support. It reads a column called cookie_string from the server table, and
sends it.
Quite simple.
I'm not too good with diff and friends, but I hope this works...
Gary (-;
cookie.diff.gz
cookie.diff.gz
I've nearly finished writing cookie support into this [but only for server
database tables; If you want to do it with indexer.conf, then you can just
use HTTPHeader directives, easy]
Does anyone want a copy?
Does anyone care?
Gary
___
If you want to
I can't get this to happen...
Can you give me an exact example of a parameter t=something, to search on,
for example,
0 and offsite
1 and andyandgary
or something?
Thank-you very much,
Gary (-;
___
If you want to unsubscribe send unsubscribe general
to
I've given up trying to get advice on XML, it's OK, I won't ask again.
Has anyone written a coherent SPROC on sybase that, given the relevant
parameters [tag, keywords, url, etc, etc] comes back with the results? In
'most any form would be ok. I mean, once someone's done the core bit...
actually, you're missing a few.
Here's the API documentation as it stands for my search-engine-on-demand
here where I work
Sorry it's in a disgusting format, but it's wahat people like at this
company.
The XML returning stuff is irrlenevant and you can ignore it. It's just
extensions I've made
Have you tried writing an independant daemon to do this?
Just send indexer SIGSTOPs whenever the load average goes above whatever,
then SIGCONTs whenever it drops again. This would have the added advantage
of being almost no resource usage, and the indexer would never need to
know...
But for
I'm outputting XML from my search engine for use in other people's websites,
and I'm having a small problem.
Some of the sites I'm indexing are made in word [I've no control over this],
and outputted as html.
And they're in strange character sets like windows-125{0,1,2}.
When I output the XML,
yep.
Just try it. I believe that the database locking should work no matter how
you do it.
And for reference, in all odds you just need to fire off more indexers on
the same box; the web server is more likely to be the bottleneck than the
box that's doing the indexing.
Hint: try running the
Has anyone here got a way of indexing powerpoint or visio documents?
Changing the document is not viable; I need a way to get the strings out of
it.
strings is not too bad on powerpoint, but for visio it's not worth the
effort.
Also, Is there any way to convert documents with this in them:
That's because what you want to do is say:
Server http://mywebstie
Alias http://mysebstie file:/path-to-files
...
Gary (-;
-Original Message-
From: Cliff Olle [SMTP:[EMAIL PROTECTED]]
Sent: Monday, April 09, 2001 12:41 AM
To: [EMAIL PROTECTED]
Subject: Problems with
OK, I've written a system that can take XML of a certain form and put it
into a crc-multi type of database.
Just have a look. I'm interested in all you feedback, especially
Alexander...
It's obviously not yet ready to put on the d/load bit of your website, but
I'm working on it...
What is this?
I'm unable to find what it is; I'm comparing some of the things in my
already existing database [generated by the indexer], and it's not the text
extract, it's not the URL itself, it's not the keywords, and it's not the
meta description.
What is it?
I'm writing an application
What is the CRC32 actually _of_?
I'm currently doing it as the CRC32 of the url itself... is this right?
And for reference, I've basically finished working on a system that takes in
XML and puts it straight into the database.
Thank-you very much,
Gary (-;
Example:
You'are searching 4 websites.
http://server1/~gbriggs
http://server2/~gbriggs
http://server3/~acoates
http://server2/~acoates
If you want to be able to search on just ~gbriggs homepages, you give them
one tag [say, "gb"].
Then you give the ~acoates homepages another tag [say, "ac"].
-Original Message-
From: Orjan Sandland [SMTP:[EMAIL PROTECTED]]
Sent: Friday, March 23, 2001 9:17 AM
To: [EMAIL PROTECTED]
Subject: Webboard: Premature end of indexing?
Author: Orjan Sandland
Email: [EMAIL PROTECTED]
Message:
I'm running latest mnogosearch and redhat
edit /etc/ld.so.conf and add /usr/local/mysql/lib to it
[don't forgfet to run /sbin/ldconfig]
or export LD_LIBRARY_PATH=/usr/local/mysql/lib:$LD_LIBRARY_PATH
or similar.
Gary (-;
-Original Message-
From: Doos [SMTP:[EMAIL PROTECTED]]
Sent: Friday, March 23, 2001 2:21 PM
To: [EMAIL
Do you want my stuff to do this?
I've wriutten some quite useful things...
Gary (-;
-Original Message-
From: Dustin S. [SMTP:[EMAIL PROTECTED]]
Sent: Sunday, March 18, 2001 8:20 AM
To: [EMAIL PROTECTED]
Subject: Will mnoGoSearch let me...
...make it so visitors can add
qtrackingalanysis.tar.gz
Hmmm.
The daemon that grabs usernames off the network; I can give that to people
if they want.
And by the way. This is based around the expanded qtracking that my system
does, it's an add-on to the php frontend that I mailed to this list a while
ago.
Gary (-;
I'll tell you what: here's _my_ personal experience.
I'm indexing many many gigabytes of data. Currently it's at ~20, but it will
go up in the near[ish] future. When I get a cluster to run it on, and the
correct access to the other files, etc.
I'm using crc-multi and most of my queries are
Why is Active in servertable an int(11)?
Surely a bool or a smallint would be faster?
Ta,
Gary (-;
___
If you want to unsubscribe send "unsubscribe general"
to [EMAIL PROTECTED]
The thing is, your URL doesn't tecnically come under the disallowed set of
things.
For that particular case, try adding /cgi/ to the Disallow path. [I'm
assuming www.pg.com/cgi is a place that you keep CGI script on your web
server]
Hope this helps,
Gary (-;
-Original Message-
From:
This one gives you the option of returning XML.
The way to use it is this:
if you use it as a regular php script
[eg http://lonwebhost20:8080/udm/blank.php ]
It does what you'd expect; returns results that're formatted based on the
template blank.htm
If, on the other hand, you run it with
010305 9:28:18 Aborted connection 394 to db: 'udmsearch' user: 'udm' host:
`localhost' (Got a packet bigger than 'max_allowed_packet')
From the mySQL error log.
Ta,
Gary (-;
___
If you want to unsubscribe send "unsubscribe general"
to [EMAIL PROTECTED]
I'm trying to lock down some of my mySQL tables, since I accidentally
deleted on of the tables the other day.
What're the minimum permissions I need to set to make searching possible?
I'm hoping to have two users:
udm [which will be used by the indexer process; password breach could be
used to
if a file is a .tar.gz file [or .tgz], you can extract it with:
gzip -cd the-file.tar.gz | tar -xf -
Gary (-;
-Original Message-
From: Mike Davis [SMTP:[EMAIL PROTECTED]]
Sent: Friday, March 02, 2001 10:29 AM
To: [EMAIL PROTECTED]
Subject: Webboard: Installation
Author:
I'm using the PHP frontend at the moment.
And I shall actually continue using it, especially when the PHP 4.0.5 comes
out (=
Gary (-;
-Original Message-
From: Laurent LEVIER [SMTP:[EMAIL PROTECTED]]
Sent: Wednesday, February 28, 2001 6:38 PM
To: [EMAIL PROTECTED]; Briggs, Gary
Hmmm.
The php one can't .
Gary (-;
-Original Message-
From: Alexander Barkov [SMTP:[EMAIL PROTECTED]]
Sent: Thursday, March 01, 2001 5:13 AM
To: [EMAIL PROTECTED]; Briggs, Gary
Subject: Re: Searching multiple tags
"Briggs, Gary" wrote:
Is there any way
For large volumes of data, which of these two is faster?
Gary (-;
-Original Message-
From: Alexander Barkov [SMTP:[EMAIL PROTECTED]]
Sent: Wednesday, February 28, 2001 9:52 AM
To: [EMAIL PROTECTED]; Briggs, Gary
Subject: Re: Crosswords and servertables
This feature works
f a series at one point]
Gary (-;
-Original Message-
From: Alexander Barkov [SMTP:[EMAIL PROTECTED]]
Sent: Wednesday, February 28, 2001 9:57 AM
To: [EMAIL PROTECTED]; Briggs, Gary
Subject: Re: Crosswords and servertables
Hi!
It seems to be a bug. CrossWeight is not working wh
Why, thank-you
Want a well-scabby patch? (=
Gary (-;
-Original Message-
From: Alexander Barkov [SMTP:[EMAIL PROTECTED]]
Sent: Wednesday, February 28, 2001 10:21 AM
To: Briggs, Gary
Cc: [EMAIL PROTECTED]
Subject: Re: Crosswords and servertables
You are right
Is there any way that I can search from multiple tags in one search?
Gary (-;
___
If you want to unsubscribe send "unsubscribe general"
to [EMAIL PROTECTED]
Is there any way that I can control what the weight of crosswords is when
all my servers are pulled out of a table on the database?
Thank-you very much,
Gary (-;
___
If you want to unsubscribe send "unsubscribe general"
to [EMAIL PROTECTED]
OK, now I feel stupid.
The PHP front end doesn't support it, right?
Does anyone know when it's likely to be supported in this frontend?
Gary (-;
-Original Message-
From: Briggs, Gary
Sent: Tuesday, February 27, 2001 2:20 PM
To: '[EMAIL PROTECTED]'; Briggs, Gary
Subject: RE
OK, It still breaks:
[lonwebhost20:/opt/udmsearch/sbin/]$ pwd
/opt/udmsearch/sbin
[lonwebhost20:/opt/udmsearch/sbin/]$ cat ./chunkystuff
#!/bin/sh
LD_LIBRARY_PATH=/opt/mySQL/lib/mysql:$LD_LIBRARY_PATH; export
LD_LIBRARY_PATH
exec ./indexer $*
[lonwebhost20:/opt/udmsearch/sbin/]$ ./chunkystuff
It's because you've got a .tar.gz file, and your version of tar [are you
using solaris, by any chance?] doesn't understand gzip compression. The way
to decompress .tar.gz files that works an almost any un*x platform is:
gzip -cd something.tar.gz | tar -xf -
Although be warned: soalris tar has a
I'm getting this too.
Solaris 2.6
MySQL 3.23.33
Mnogo 3.1.11
compiled with
CC="/opt/SUNWspro/bin/cc";export CC
CFLAGS="-fast";export CFLAGS
CXX="/opt/SUNWspro/bin/CC";export CXX
./configure --cache-file=/dev/null --prefix=/opt/udmsearch \
--with-mysql=/opt/mySQL --disable-syslog
OK, so I've been reading the source code, and I'm having real trouble with
what some parts of the url table are for...
Why bother to compute a crc32 for the urls? do I need it? [I'm currently
using crc-multi db mode on mySQL]
It seems to be the primary key, but then what's the point of rec_id?
41 matches
Mail list logo