I am looking at trying to use ferret/aaf to supplement my querying against a medium and large table with lots of columns. Some facts first:
Ferret 0.11.4 AAF 0.4.0 Ruby 1.8.6 Rails 1.2.3 Medium table: 105,464 rows 168 columns (mostly varchar(20)) 11 actual columns indexed in aaf plus 40 virtual columns indexed in aaf (virtual is concat of two physical columns. e.g. cast_first_name_1 + cast_last_name_1 through cast_first_name_20 + cast_last_name_20) Large table: 1,244,716 rows same column/index structure These tables are not updated via Ruby, only read. I am trying to use rebuild_index to bootstrap the medium sized table and it is taking a very long time (running for about 4 hours, indicates 50% complete with 4 hours remaining) and creating a massive number of files in the index directory (currently about 65k, was 90k earlier) I have not done any tuning of ferret/aaf so far, and I fear what it will look like to do the big table. Does anyone have any advise on how to speed this process up? Because the tables are updated by an external batch process, if I were to continue down this ferret/aaf path, I'd have to be looking at running this rebuild_index a couple of times per week which would be rather painful given the present time and might not be possible if the large table took more than 48 hours... _______________________________________________ Ferret-talk mailing list [email protected] http://rubyforge.org/mailman/listinfo/ferret-talk

