Hi,

I am working on deeply nested directory structure and large logs in perl and
trying to generate the results from index data,

my $hits = $searcher->hits(
        query      => ['title'],
        num_wanted => -1,
    );

while ( my $hit = $hits->next ) {
    # making 28777 calls to Lucy::Search::Hits::next 
    # do some work - already ran profiling on this code and optimized 

Now, since the hit count is large, the stuff that is in the iteration is
consuming good amount of time, i need to improve its performance to get it
going, 

can we run this in parallel or any other optimization possible for case
where hits are present in thousands? 

I tried implementing Parallel::ForkManager in it, but it had increased the
running time to serious extent instead of reducing it, 

Not getting any clue, please help i am badly stuck now.

Regards
Rohit Singh





--
View this message in context: 
http://lucene.472066.n3.nabble.com/iterating-through-hits-is-there-a-way-to-improve-performance-or-can-we-run-these-iterations-in-parall-tp4329286.html
Sent from the lucy-user mailing list archive at Nabble.com.

Reply via email to