Still looking for answer on Schema.xml and SolrConfig.xml

1.       Do I need to tell Solr, to extract Title from PDF, go look for Title 
word and extract entire line after the Tag and collect all such occurrence’s 
from hundreds of PDFs and build the Title column data and index it?


2.       How to define my own schema to Solr

3.       Say I defined my fields Title, Ticket_number, Submitter, client and so 
on, How can I verify respective data is extracted in specific columns in Solr 
and indexed? Any suggestion on how the Analyzer, Tokenizer and Filter and which 
one will help for this purpose?


1.       I do not want to dump entire 4 GB PDF contents in one searchable field 
(ATTR_CONTENT) in Solr

2.       Even if entire PDF contents is extracted in above field as a default, 
I still want to extract specific searchable column data in their respective 
fields

3.       Rather I want to configure Solr to have column wise searchable 
contents such as Title, number, and so on

Any suggestions on performance? PDF database is 80 GB, will it be fast enough? 
Do I Need to divide in multiple cores and on multiple machines ? and on 
multiple web apps? And clustering?


I should have mentioned my PDFs are from Ticketing system like Jira which is 
already retired way back from production and all I have is the Ticketing system 
PDF database.


4.       My system will be used internally just by the selected number of very 
few people.

5.       They can wait 4 GB PDF to get loaded.

6.       I agree there will be many matches will be found in one large PDF, 
based on search criteria

7.       To make searches faster I want Solr to create more columns and column 
based indexes

8.       Solr underneath uses Tika which is extracting contents and getting rid 
of all the rich content formatting characters present in the PDF document.

9.       I believe resulting extraction size is 1/5th of the original PDF 
..just a random guess based on one sample extraction




From: "Jürgen Wagner (DVT)" [mailto:juergen.wag...@devoteam.com]
Sent: Tuesday, January 06, 2015 11:56 AM
To: solr-user@lucene.apache.org
Subject: Re: PDF search functionality using Solr

Hello,
  no matter which search platform you will use, this will pose two challenges:

- The size of the documents will render search less and less useful as the 
likelihood of matches increases with document size. So, without a proper 
semantic extraction (e.g., using decent NER or relationship extraction with a 
commercial text mining product), I doubt you will get the required precision to 
make this overly usefiul.

- PDFs can have their own character sets based on the characters actually used. 
Such file-specific character sets are almost impossible to parse, i.e., if your 
PDFs happen to use this "feature" of the PDF format, you won't be lucky getting 
any meaningful text out of them.

My suggestion is to use the Jira REST API to collect all necessary documents 
and index the resulting XML or attachment formats. As the REST API provides 
filtering capabilities, you could easily create incremental feeds to avoid 
humongous indexing every time there's new information in Jira. Dumping Jira 
stuff as PDF seems to me to be the least suitable way of handling this.

Best regards,
--Jürgen


On 06.01.2015 18:30, ganesh.ya...@sungard.com<mailto:ganesh.ya...@sungard.com> 
wrote:

Hello Solr-users and developers,

Can you please suggest,



1.       What I should do to index PDF content information column wise?



2.       Do I need to extract the contents using one of the Analyzer, Tokenize 
and Filter combination and then add it to Index? How can test the results on 
command prompt? I do not know the selection of specific Analyzer, Tokenizer and 
Filter for this purpose



3.       How can I verify that the needed column info is extracted out of PDF 
and is indexed?



4.       So for example How to verify Ticket number is extracted in 
Ticket_number tag and is indexed?



5.       Is it ok to post 4 GB worth of PDF to be imported and indexed by Solr? 
I think I saw some posts complaining on how large size that can be posted ?



6.       What will enable Solr to search in any PDF out of many, with different 
words such as "Runtime" "Error" "XXXX" and result will provide the link to the 
PDF



My PDFs are nothing but Jira ticket system.

PDF has info on

Ticket Number:

Desc:

Client:

Status:

Submitter:

And so on:





1.       I imported PDF document in Solr and it does the necessary searching 
and I can test some of it using the browse client interface provided.



2.       I have 80 GB worth of PDFs.



3.       Total number of PDFs are about 200



4.       Many PDFs are of size 4 GB



5.       What do you suggest me to import such a large PDFs? What tools can you 
suggest to extract PDF contents first in some XML format and later Post that 
XML to be indexed by Solr.?















Your early response is much appreciated.







Thanks



G





--

Mit freundlichen Grüßen/Kind regards/Cordialement vôtre/Atentamente/С уважением 
i.A. Jürgen Wagner Head of Competence Center "Intelligence"
& Senior Cloud Consultant

Devoteam GmbH, Industriestr. 3, 70565 Stuttgart, Germany
Phone: +49 6151 868-8725, Fax: +49 711 13353-53, Mobile: +49 171 864 1543
E-Mail: juergen.wag...@devoteam.com<mailto:juergen.wag...@devoteam.com>, URL: 
www.devoteam.de<http://www.devoteam.de/>

________________________________
Managing Board: Jürgen Hatzipantelis (CEO) Address of Record: 64331 
Weiterstadt, Germany; Commercial Register: Amtsgericht Darmstadt HRB 6450; Tax 
Number: DE 172 993 071


Reply via email to