Hello,

A few minutes ago we have uploaded aspseek-1.2.6 sources
to SW sites. Find excerpts from NEWS and RELEASE-NOTES below,
download from http://www.aspseek.org/ as usual.

Binary RPMS will be available in a few days.

Enjoy,
   Kir.



from NEWS:
-----
14 Nov 2001: v.1.2.6 (stable)                                                   
                                                                                
* Changed logs.txt format
* Implemented buddy like heap for storing of position vectors in "index"
  for better memory usage
* Fixed improper clones processing in "index"
* Implemented buffered file to use instead of FILE* during inverted index
  merging for faster processing
* Fixed closing of unnecessary pipe handles in resolver and parent "index"
  processes
* Fixed coredumps in "searchd" which occured when "Cache" mode is on and
  complex expressions are searched
* Added UrlBufferSize parameter to aspseek.conf
* URL IDs sorting is performed by STL sort function instead of "ORDER BY"
  during subsets generation
* Changed output format of "index"
* MaxDocsAtOnce is ignored now for sites to which connection was failed
  last time
* Fixed rare coredump in "searchd" if "site:" was used in query
* Reduced amount of stack used by "index" to optimize memory usage and
  eliminate rare coredumps occured during citations merging
* Added MultipleDBConnections parameter to searchd.conf
* Time of SEGV in "searchd" is stored in dlog.log now
* Fixed not important memory leak in "index" which occured during ranks
  calculation
* Added -X and -H flags to "index"
* SIGCHLD signals are caught and logged by "index"
* Fixed huge memory leak in "SetNewOrigin" function of "index" which
  could occur only when index was run to re-index existing database
* Fixed rare coredump in "SetNewOrigin" function of "index"
* Event of breaking of read pipe in resolver process is logged now
* Exiting of resolver process is logged now
* Optimized performance of "searchd" by replacing of many small "read"
  calls to less number of big "read" calls in CResult::GetUrls(CSiteInfo*)
* Fixed -P and -A flags in "index"
* Optimized memory usage by reallocating of URL content buffer which
  was filled by more than 100000 bytes last time
* Removed MaxNetErrors parameter from aspseek.conf (as it is unused)
* Added description of DeltaBufferSize, UrlBufferSize, WordCacheSize,
  HrefCacheSize, NextDocLimit, UtfStorage, MaxDocsAtOnce,
  MultipleDBConnections parameters to man pages
* Added RELEASE-NOTES to the distribution
-----

from RELEASE-NOTES:
-----
        v.1.2.6

In this version we have fixed numerous small and not-so-small bugs, as well
as added some improvements and new features. Here some changes are described
in details; for full list of changes see the NEWS file.

To summarize, this version should be more stable, as we have found the
reasons and fixed a few coredump cases, as well as healed some memory
leaks. Also, some tricks were incorporated to boost performance: buddy heap
is implemented for storing word position vectors in "index", replacing the
standard STL vector class; CBufferedFile is implemented and is used instead
of stdio during inverted index merging process (also known as "Saving
delta files").

Some other items from NEWS is explained here.

* Added MultipleDBConnections parameter to searchd.conf

Multiple DB connections feature is implemented in searchd. This improves
concurrency between searchd threads, especially in the case when some threads
do pattern search (which is slow), and other thread wants a simple search.

* Exiting of resolver process is logged now
* Event of breaking of read pipe in resolver process is logged now
* SIGCHLD signals are caught by "index"

We have some reports from users of 1.2.5 with complains about stuck of
index process. The problem seems to be in the resolver process. So, we have
added several facilities that can be helpful to nail down the bug.

* Added -X and -H flags to "index"
Use "index -X1" to check inverted index for URLs for which "urlword.deleted"
field is non-zero. Use "index -X2" to fix it by appending information about
deleted keys to the delta files. So if you want to remove records where
"urlword.deleted" is non-zero, run index -X2; index -D, and finally perform
SQL statements to delete unnecessary records.

"index -H" is used to recreate citation indexes and ranks file from
"urlwordsNN.hrefs" fields in case of citation index corruption.

* Fixed -P and -A flags in "index"

"index -P" is used to show "path" to given URL.

"index -A" is used to add or delete site to/from web space.
-----

-- 
[EMAIL PROTECTED]  ICQ 7551596  Phone +7 903 6722750
Hard work may not kill you,  but why take chances?
--

Reply via email to