Re: Storing MYSQL DATETIME field in solr as String
OK here are the scenarios I tried. *Scenario - 1: * dih.xml (aka data-config.xml) entity dataSource=solr name=listing query=... transformer=DateFormatTransformer field name=quot;publish_datequot; column=quot;publish_datequot; xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd'T'HH:mm:ss* locale=en/ schema.xml field name=publish_date type=date indexed=true stored=true multiValued=false default=NOW / It does not work. Throws an following exception WARN org.apache.solr.handler.dataimport.DateFormatTransformer – Could not parse a Date field java.text.ParseException: Unparseable date: 2013-12-05 15:40:03.0 *Scenario - 2* dih.xml (aka data-config.xml) entity dataSource=solr name=listing query=... transformer=DateFormatTransformer field name=quot;publish_datequot; column=quot;publish_datequot; xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd HH:mm:ss* locale=en/ schema.xml field name=publish_date type=date indexed=true stored=true multiValued=false default=NOW / Output is date name=publish_date2012-12-05T07:38:27Z/date *Scenario - 3* dih.xml (aka data-config.xml) entity dataSource=solr name=listing query=... transformer=DateFormatTransformer field name=quot;publish_datequot; column=quot;publish_datequot; xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd HH:mm:ss* locale=en/ schema.xml field name=publish_date type=string indexed=true stored=true multiValued=false default=NOW / Output is str name=publish_dateThu Dec 05 15:38:27 MYT 2013/str My expectation was *2012-12-05 07:38:27* (MySQL DateTime format) -- View this message in context: http://lucene.472066.n3.nabble.com/Storing-MYSQL-DATETIME-field-in-solr-as-String-tp4106836p4106854.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Solr Profiler
Hi NewRelic is good enough to monitor the Solr. Are you using Solarium or SolrJ client to connect to Solr?. We have used Solarium and able to monitor each calls and gather most of the info. -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Profiler-tp4106231p4106859.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Backwards compatibility issue
Thanks for your feedback. Upgrading SolrJ is going to be a bit difficult (because it's tied in with a framework which is only partially owned by us), so we'll just have to find another way to deal with it. Karel On Fri, Dec 13, 2013 at 7:28 PM, Shawn Heisey s...@elyograg.org wrote: On 12/13/2013 7:40 AM, Karel Vervaeke wrote: IIUC A Solrj 4.4 client should be able to speak to a Solr 4.6 cloud instance. Assuming this is correct there's an issue because of a change in clusterstate.json: In a 4.4 cloud server clusterstate.json would contain `router: implicit` In a 4.6 cloud server clusterstate.json contains `router: {name:implicit}`. Because of this a 4.4 solrj client will look up a router (in DocRouter.java) using the key `{name:implicit}` instead of just `implicit`. Is there a way around this that does not involve upgrading the solrj client? Although the committers do try really really hard to maintain cross-version compatibility between SolrJ and Solr versions, sometimes changes must be made that break that compatibility, either to support critical new functionality or because of bugs or situations that were not anticipated. SolrCloud technology is relatively new, so things are changing quickly to accommodate new functionality and fix stability problems. SOLR-4221 made this change (released in version 4.5.0) so that more than one property could be assigned to 'router' because more information is sometimes needed. https://issues.apache.org/jira/browse/SOLR-4221 One of the commits made on this issue was to allow a newer SolrJ to talk to an older cluster with the previous clusterstate router format. If for some reason you need to run different SolrJ and Solr versions, it is almost always better to have a newer SolrJ than Sol Thanks, Shawn
Heap size and Solr 4.3
Hi All, we have deployed on our production environment a new Solr 4.3 instance (2 nodes with SolrCloud) but this morning one node gone on outofmemory status and we have noticed that the JVM uses a lot of Old Gen space during the normal lifecycle. What are the items that improve this high usage of Heap? Thanks, Marcello
Re: Poor performance on distributed search
Yonik Seeley-2-2 wrote On Wed, Dec 28, 2011 at 5:47 AM, ku3ia lt; demesg@ gt; wrote: So, based on p.2) and on my previous researches, I conclude, that the more documents I want to retrieve, the slower is search and main problem is the cycle in writeDocs method. Am I right? Can you advice something in this situation? For the first phase in a distributed search, Solr must return the top N ids (in your case 200). It currently does this by loading stored fields, which is slow. A better approach is to store the id field as a column stride field. https://issues.apache.org/jira/browse/SOLR-2753 -Yonik http://www.lucidimagination.com Hi all again. I have upgraded Solr version to 4.6.0 and want to test DocValues fields in action. My schema is: 4 shards of 48M of docs each. All shards are on the same machine. I have turned on DV for id field in my schema.xml: field name=RecordID type=uid indexed=true stored=true required=true docValues=true/ ... uniqueKeyRecordID/uniqueKey ... fieldType name=uid class=solr.StrField sortMissingLast=true docValuesFormat=Disk/ I had created 30 queries like, q=(a OR b) OR (c OR d e OR f), where a through f are random words These queries were curled each other, not concurrently. After that I turned off doc values, rebuild all four shards and runs queries again. At results, a haven't seen any difference between stored ID field and docValues and one more moment, when I set stored=false and docValues=true for id field I had an error when running distributed search: ERROR - 2013-12-12 16:59:45.508; org.apache.solr.common.SolrException; java.lang.NullPointerException at org.apache.solr.handler.component.QueryComponent.mergeIds(QueryComponent.java:1004) at org.apache.solr.handler.component.QueryComponent.handleRegularResponses(QueryComponent.java:750) at org.apache.solr.handler.component.QueryComponent.handleResponses(QueryComponent.java:729) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:312) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1863) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:710) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:413) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419) at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137) at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557) at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231) at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075) at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384) at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193) at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135) at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255) at org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116) at org.eclipse.jetty.server.Server.handle(Server.java:370) at org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489) at org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:949) at org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1011) at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:644) at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235) at org.eclipse.jetty.server.AsyncHttpConnection.handle(AsyncHttpConnection.java:82) at org.eclipse.jetty.io.nio.SelectChannelEndPoint.handle(SelectChannelEndPoint.java:668) at org.eclipse.jetty.io.nio.SelectChannelEndPoint$1.run(SelectChannelEndPoint.java:52) at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:608) at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:543) at java.lang.Thread.run(Thread.java:744) This error is on QueryComponent on merge Ids: resultIds.put(shardDoc.id.toString(), shardDoc); So seems it try to use stored id, not docvalue? I tried different codecs: Direct, Disk, Memory, Lucene45. Results are nearby the
OutOfMemoryError in RamUsageEstimator in solr 4.6
Hi, we get a OutOfMemoryError in RamUsageEstimator and are a little bit confused about the error. We are using solr 4.6 and are confused about the Lucene42DocValuesProducer. We checked current solr code and found that Lucene42NormsFormat will be returned as NormFormat in Lucene46Codec and so the Lucene42DocValuesProducer will be returned. Is there any special reason not to have a higher version of NormsFormat? This is our StackTrace: SolrException|null:java.lang.OutOfMemoryError: Java heap space at org.apache.lucene.util.RamUsageEstimator$IdentityHashSet.allocateBuffers(RamUsageEstimator.java:762) at org.apache.lucene.util.RamUsageEstimator$IdentityHashSet.expandAndRehash(RamUsageEstimator.java:736) at org.apache.lucene.util.RamUsageEstimator$IdentityHashSet.add(RamUsageEstimator.java:678) at org.apache.lucene.util.RamUsageEstimator.measureObjectSize(RamUsageEstimator.java:437) at org.apache.lucene.util.RamUsageEstimator.sizeOf(RamUsageEstimator.java:350) at org.apache.lucene.codecs.lucene42.Lucene42DocValuesProducer.ramBytesUsed(Lucene42DocValuesProducer.java:194) at org.apache.lucene.index.SegmentCoreReaders.ramBytesUsed(SegmentCoreReaders.java:195) at org.apache.lucene.index.SegmentReader.ramBytesUsed(SegmentReader.java:558) at org.apache.solr.handler.admin.LukeRequestHandler.getIndexHeapUsed(LukeRequestHandler.java:579) at org.apache.solr.handler.admin.LukeRequestHandler.getIndexInfo(LukeRequestHandler.java:558) at org.apache.solr.handler.admin.CoreAdminHandler.getCoreStatus(CoreAdminHandler.java:1044) at org.apache.solr.handler.admin.CoreAdminHandler.handleStatusAction(CoreAdminHandler.java:685) at org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:167) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) at org.apache.solr.servlet.SolrDispatchFilter.handleAdminRequest(SolrDispatchFilter.java:662) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:248) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:243) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:210) ... Please contact me, if there are further questions. Best regards Torben Greulich -- Torben Greulich · Entwickler Backend Development shopping24 internet group Wandsbeker Straße 3-7 · 22172 Hamburg Telefon: +49 (0) 40 6461 · Fax: +49 (0) 40 6461 7879 torben.greul...@s24.com · www.s24.com AG Hamburg HRB 63371 vertreten durch Dr. Björn Schäfers und Martin Mildner
How to boost documents ?
Hi, How to boost documents that contain all search terms in several of its fields ? Below you cand find a simplified example : The query with Min should match: q=beautiful Christmas treemm=2qf=title^12 description^2 There are two offers that match the query : offer1 {title:Christmas tree, description:a joy for children} offer2 {title:Christmas tree, description:beautiful for holidays}} The first offer ranks before the second, despite of the fact that the second one contains all the search terms. I tried to play with the boosts of qf, but the results vary a lot. Is there a way to add a boost on all search fields, the same way we do with pf on one field : pf=title:2^3.0 ? Thank you, Anca -- Anca Kopetz Software engineer E anca.kop...@kelkoo.comY!Messenger kelkooancak T 33 (0)4 56 09 07 55 A 4/6 Rue des Meridiens 38130 Echirolles Kelkoo SAS Société par Actions Simplifiée Au capital de € 4.168.964,30 Siège social : 8, rue du Sentier 75002 Paris 425 093 069 RCS Paris Ce message et les pièces jointes sont confidentiels et établis à l'attention exclusive de leurs destinataires. Si vous n'êtes pas le destinataire de ce message, merci de le détruire et d'en avertir l'expéditeur.
indexing from bowser
how to index pdf,doc files from browser? this query is used for indexing : curl http://localhost:8080/solr/document/update/extract?literal.id=12commit=true; -Fmyfile=@C:\solr\document\src\test1\Coding.pdf but i need to index from browser, as we do for delete: http://localhost:8080/solr/document/update?stream.body=deletequeryid:3/query/deletecommit=true When i try to index using this: http://localhost:8080/solr/document/update/extract?literal.id=12commit=true; -Fmyfile=@C:\solr\document\src\test1\Coding.pdf the document does not get indexed. Is there any query to index from browser,using html url?? -- View this message in context: http://lucene.472066.n3.nabble.com/indexing-from-bowser-tp4106889.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: indexing from bowser
Hi, (13/12/16 19:46), Nutan wrote: how to index pdf,doc files from browser? I think you can index from browser. If you said that this query is used for indexing : curl http://localhost:8080/solr/document/update/extract?literal.id=12commit=true; -Fmyfile=@C:\solr\document\src\test1\Coding.pdf curl works for you but When i try to index using this: http://localhost:8080/solr/document/update/extract?literal.id=12commit=true; -Fmyfile=@C:\solr\document\src\test1\Coding.pdf the document does not get indexed. browser doesn't work for you, why don't you look into Solr log and compare the logs between when you using curl and browser? koji -- http://soleami.com/blog/automatically-acquiring-synonym-knowledge-from-wikipedia.html
Re: indexing from bowser
On 16 December 2013 16:30, Koji Sekiguchi k...@r.email.ne.jp wrote: Hi, (13/12/16 19:46), Nutan wrote: how to index pdf,doc files from browser? I think you can index from browser. If you said that this query is used for indexing : curl http://localhost:8080/solr/document/update/extract?literal.id=12commit=true; -Fmyfile=@C:\solr\document\src\test1\Coding.pdf curl works for you but When i try to index using this: http://localhost:8080/solr/document/update/extract?literal.id=12commit=true; -Fmyfile=@C:\solr\document\src\test1\Coding.pdf the document does not get indexed. browser doesn't work for you, why don't you look into Solr log and compare the logs between when you using curl and browser? No, this will not work by loading the URL into a browser, as with the -F option, curl is emulating a POST for a form filled in by the user. One must use curl, or a similar tool, for this. Regards, Gora
High load from solr 4.5
I tried to install some versions of solr, from solr 4 to solr 4.6 (the last release) in my production environment. The configuration of the machine is - Debian 7 64bit - java opendjk 7 - CPU 8 core - 16Gb ram but i notice that the machine load grows if i install the version 4.4 or above. Why? the configurations (solrconfig and schema) are the same. The indexes are the same. I don't use solrcloud. I run only one instance and i give to solr 6 gb or ram. The logs are disable. What the difference? thank you -- *Gastone Penzo*
Re: indexing from bowser
ok thanks, but is there any other way where -F is not used? I am creating a api in vc++ and to link to solr i am using libcurl,for this to work the string is the url, eg: curl_easy_setopt(curl, CURLOPT_URL,http://localhost:8080/solr/document/select?q=*%3A*wt=jsonindent=truefl=id;); so for indexing i need a url. -- View this message in context: http://lucene.472066.n3.nabble.com/indexing-from-bowser-tp4106889p4106900.html Sent from the Solr - User mailing list archive at Nabble.com.
RE: SOLR DIH - Sub Entity with different datasource not working
Thanks for the reply. The error shows it is not able to execute the query. In my case, if you see my config file I am joining the entities between two different datasources.. i.e., Entity1 - Datasource1 --Subentity - DataSource2 My doubt is, can we join the entities in two different datasources. Is this fine.. -- View this message in context: http://lucene.472066.n3.nabble.com/SOLR-DIH-Sub-Entity-with-different-datasource-not-working-tp4106550p4106906.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SOLR DIH - Sub Entity with different datasource not working
hi i dont know if this can be done but to avoid this you can create a new table with the results and index that new table :) you can then delete the table as well after indexing ... :) tc cheers karan On Mon, Dec 16, 2013 at 5:42 PM, Lokn nlokesh...@gmail.com wrote: Thanks for the reply. The error shows it is not able to execute the query. In my case, if you see my config file I am joining the entities between two different datasources.. i.e., Entity1 - Datasource1 --Subentity - DataSource2 My doubt is, can we join the entities in two different datasources. Is this fine.. -- View this message in context: http://lucene.472066.n3.nabble.com/SOLR-DIH-Sub-Entity-with-different-datasource-not-working-tp4106550p4106906.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: How to boost documents ?
Hi Anca, Can you try following URL? q=beautiful Christmas treemm=2qf=title^12 description^2defType=dismaxbf=map(query($qq),0,0,0,100.0)qq={!dismax qf='title description' mm=100%}beautiful Christmas tree Modified from Jan's solution. See his original post [1] to a similar discussion. [1] http://search-lucene.com/m/nK6t9j1fuc2 On Monday, December 16, 2013 12:19 PM, Anca Kopetz anca.kop...@kelkoo.com wrote: Hi, How to boost documents that contain all search terms in several of its fields ? Below you cand find a simplified example : The query with Min should match: q=beautiful Christmas treemm=2qf=title^12 description^2 There are two offers that match the query : offer1 {title:Christmas tree, description:a joy for children} offer2 {title:Christmas tree, description:beautiful for holidays}} The first offer ranks before the second, despite of the fact that the second one contains all the search terms. I tried to play with the boosts of qf, but the results vary a lot. Is there a way to add a boost on all search fields, the same way we do with pf on one field : pf=title:2^3.0 ? Thank you, Anca -- Anca Kopetz Software engineer eanca.kop...@kelkoo.com Y!Messenger kelkooancak T +33 (0)4 56 09 07 55 A 4/6 Rue des Meridiens 38130 Echirolles Kelkoo SAS Société par Actions Simplifiée Au capital de € 4.168.964,30 Siège social : 8, rue du Sentier 75002 Paris 425 093 069 RCS Paris Ce message et les pièces jointes sont confidentiels et établis à l'attention exclusive de leurs destinataires. Si vous n'êtes pas le destinataire de ce message, merci de le détruire et d'en avertir l'expéditeur.
Re: uniquekey generation in solr
one of the field used in unique key generation is in date format and other are in string format. So when we remove this date field ,it is working fine. -- View this message in context: http://lucene.472066.n3.nabble.com/uniquekey-generation-in-solr-tp4106766p4106919.html Sent from the Solr - User mailing list archive at Nabble.com.
External File Field
Hi Is there a way to find ,whether the External File Fields mentioned in the schema.xml is being used or whether Solr reads the value of those external fields. I am not sure how to use an External field, can I request the value of an External File Field in the field list or can I use in my custom function as shown below: http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,adjlocalityhttp://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,property_price_min,lanorm(%22%22,luxury_amenities_c,10),adjlocality ,custfunc(adjlocality). I didn't get any output for the external fields in the queries that I provide. Thanks Regards Mukund
Re: External File Field
Hi, You can request the value of the that field via fl=*,field(adjlocality) See more about it : https://cwiki.apache.org/confluence/display/solr/Working+with+External+Files+and+Processes Actually you can search it too with frange query parser. {!frange l=0 u=0}field(adjlocality) On Monday, December 16, 2013 5:05 PM, Mukundaraman valakumaresan muk...@8kmiles.com wrote: Hi Is there a way to find ,whether the External File Fields mentioned in the schema.xml is being used or whether Solr reads the value of those external fields. I am not sure how to use an External field, can I request the value of an External File Field in the field list or can I use in my custom function as shown below: http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,adjlocalityhttp://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,property_price_min,lanorm(%22%22,luxury_amenities_c,10),adjlocality ,custfunc(adjlocality). I didn't get any output for the external fields in the queries that I provide. Thanks Regards Mukund
RE: How can you move a shard from one SolrCloud node to another?
Hi Chris, The easiest approach is to just create a new core on the new machine that references the collection and shard you want to migrate. For example, say you split shard1 of a collection named cloud, which results in having: shard1_0 and shard1_1. Now let's say you want to migrate shard 1_0 over to the new machine. First, fire off a q=*:*distrib=false query to the shard you're migrating so that you know how many docs it has (which will be used to verify the migration was clean below). Next, bring up the new machine in cloud mode (-zkHost=?) and then go to the admin console on that server. Nav to the core admin page and create a new core, specifying the collection and shard1_0 in the form; note: the form leads you to believe you need to create the directory on the local system but you actually don't need to worry about doing that as the config will get pulled from ZK and the directory will get created on the fly (at least that's what happened in my env using branch_4x). When the new core initializes, it will use good ol' snapshot replication to pull the index from the leader. Verify the new core is happy by executing the q=*:*distrib=false query again. Once you're satisfied, you can unload the core you migrated. Btw ... you can do all this with the core admin API instead of the Web UI if you want to script it. Cheers, Timothy Potter Sr. Software Engineer, LucidWorks www.lucidworks.com From: cwhi chris.whi...@gmail.com Sent: Sunday, December 15, 2013 3:43 PM To: solr-user@lucene.apache.org Subject: How can you move a shard from one SolrCloud node to another? Let's say I want to rebalance a SolrCloud collection. I call SPLITSHARD to split an existing shard, and then I'd like to move one of the subshards to a new machine so the index is more balanced. Can this be done? If not, how do you rebalance an existing SolrCloud collection? -- View this message in context: http://lucene.472066.n3.nabble.com/How-can-you-move-a-shard-from-one-SolrCloud-node-to-another-tp4106815.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: How can you move a shard from one SolrCloud node to another?
Tim, Can you explain how the replication snapshot is done using the coreAdminAPI? -- Yago Riveiro Sent with Sparrow (http://www.sparrowmailapp.com/?sig) On Monday, December 16, 2013 at 4:23 PM, Tim Potter wrote: Hi Chris, The easiest approach is to just create a new core on the new machine that references the collection and shard you want to migrate. For example, say you split shard1 of a collection named cloud, which results in having: shard1_0 and shard1_1. Now let's say you want to migrate shard 1_0 over to the new machine. First, fire off a q=*:*distrib=false query to the shard you're migrating so that you know how many docs it has (which will be used to verify the migration was clean below). Next, bring up the new machine in cloud mode (-zkHost=?) and then go to the admin console on that server. Nav to the core admin page and create a new core, specifying the collection and shard1_0 in the form; note: the form leads you to believe you need to create the directory on the local system but you actually don't need to worry about doing that as the config will get pulled from ZK and the directory will get created on the fly (at least that's what happened in my env using branch_4x). When the new core initializes, it will use good ol' snapshot replication to pull the index from the leader. Verify the new core is happy by executing the q=*:*distrib=false query again. Once you're satisfied, you can unload the core you migrated. Btw ... you can do all this with the core admin API instead of the Web UI if you want to script it. Cheers, Timothy Potter Sr. Software Engineer, LucidWorks www.lucidworks.com (http://www.lucidworks.com) From: cwhi chris.whi...@gmail.com (mailto:chris.whi...@gmail.com) Sent: Sunday, December 15, 2013 3:43 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: How can you move a shard from one SolrCloud node to another? Let's say I want to rebalance a SolrCloud collection. I call SPLITSHARD to split an existing shard, and then I'd like to move one of the subshards to a new machine so the index is more balanced. Can this be done? If not, how do you rebalance an existing SolrCloud collection? -- View this message in context: http://lucene.472066.n3.nabble.com/How-can-you-move-a-shard-from-one-SolrCloud-node-to-another-tp4106815.html Sent from the Solr - User mailing list archive at Nabble.com (http://Nabble.com).
Re: indexing from bowser
On 16 December 2013 16:50, Nutan nutanshinde1...@gmail.com wrote: ok thanks, but is there any other way where -F is not used? I am creating a api in vc++ and to link to solr i am using libcurl,for this to work the string is the url, eg: curl_easy_setopt(curl, CURLOPT_URL,http://localhost:8080/solr/document/select?q=*%3A*wt=jsonindent=truefl=id;); [...] If you can do a post using curl from the command line, you can add --libcurl file to get a libcurl example written to file. Please see the curl man page for details. Regards, Gora
Re: OutOfMemoryError in RamUsageEstimator in solr 4.6
On 12/16/2013 2:34 AM, Torben Greulich wrote: we get a OutOfMemoryError in RamUsageEstimator and are a little bit confused about the error. We are using solr 4.6 and are confused about the Lucene42DocValuesProducer. We checked current solr code and found that Lucene42NormsFormat will be returned as NormFormat in Lucene46Codec and so the Lucene42DocValuesProducer will be returned. Is there any special reason not to have a higher version of NormsFormat? The format for that part of the index apparently hasn't changed since Lucene 4.2. Something else must have changed about the default index format, or I expect that there would not be a 4.6 specific Codec object. From what I've seen, the verion number encoded in the class name doesn't change until there's a new class needed for the default codec. A java OutOfMemoryError means that you need to increase the max heap when you start the program. It looks like you're trying to get a core status, so this is probably happening when using the admin UI. To put it quite simply, your Java max heap is not big enough to handle what Solr has been asked to do. http://wiki.apache.org/solr/SolrPerformanceProblems#Java_Heap Thanks, Shawn
Re: Poor performance on distributed search
Any ideas? -- View this message in context: http://lucene.472066.n3.nabble.com/Poor-performance-on-distributed-search-tp3590028p4106968.html Sent from the Solr - User mailing list archive at Nabble.com.
RE: How can you move a shard from one SolrCloud node to another?
Hi Yago, When you create a new core (via API or Web UI), you specify the collection name and shard id, in my example cloud and shard1_0. When the core initializes in SolrCloud mode, it recognizes that the collection exists and adds itself as a replica to the shard. Then the main replica recovery process kicks in; try PeerSync, realize too far out of date, try snapshot replication from leader. The following core API command led to the same result as using the UI: curl -v http://localhost:8986/solr/admin/cores?action=CREATEcollection=cloudshard=shard1_0name=cloud_shard1_0_replica3; The only trick here is you need to set the name of the core, which from what I can tell can be arbitrary but I chose to use the same naming standard as the other cores Cheers, Timothy Potter Sr. Software Engineer, LucidWorks www.lucidworks.com From: Yago Riveiro yago.rive...@gmail.com Sent: Monday, December 16, 2013 9:32 AM To: solr-user@lucene.apache.org Subject: Re: How can you move a shard from one SolrCloud node to another? Tim, Can you explain how the replication snapshot is done using the coreAdminAPI? -- Yago Riveiro Sent with Sparrow (http://www.sparrowmailapp.com/?sig) On Monday, December 16, 2013 at 4:23 PM, Tim Potter wrote: Hi Chris, The easiest approach is to just create a new core on the new machine that references the collection and shard you want to migrate. For example, say you split shard1 of a collection named cloud, which results in having: shard1_0 and shard1_1. Now let's say you want to migrate shard 1_0 over to the new machine. First, fire off a q=*:*distrib=false query to the shard you're migrating so that you know how many docs it has (which will be used to verify the migration was clean below). Next, bring up the new machine in cloud mode (-zkHost=?) and then go to the admin console on that server. Nav to the core admin page and create a new core, specifying the collection and shard1_0 in the form; note: the form leads you to believe you need to create the directory on the local system but you actually don't need to worry about doing that as the config will get pulled from ZK and the directory will get created on the fly (at least that's what happened in my env using branch_4x). When the new core initializes, it will use good ol' snapshot replication to pull the index from the leader. Verify the new core is happy by executing the q=*:*distrib=false query again. Once you're satisfied, you can unload the core you migrated. Btw ... you can do all this with the core admin API instead of the Web UI if you want to script it. Cheers, Timothy Potter Sr. Software Engineer, LucidWorks www.lucidworks.com (http://www.lucidworks.com) From: cwhi chris.whi...@gmail.com (mailto:chris.whi...@gmail.com) Sent: Sunday, December 15, 2013 3:43 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: How can you move a shard from one SolrCloud node to another? Let's say I want to rebalance a SolrCloud collection. I call SPLITSHARD to split an existing shard, and then I'd like to move one of the subshards to a new machine so the index is more balanced. Can this be done? If not, how do you rebalance an existing SolrCloud collection? -- View this message in context: http://lucene.472066.n3.nabble.com/How-can-you-move-a-shard-from-one-SolrCloud-node-to-another-tp4106815.html Sent from the Solr - User mailing list archive at Nabble.com (http://Nabble.com).
RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup
Also, I'm aware that there's two typos in my schema.xml attached. I forgot to remove the linebreak \ character from the two splitOnCaseChange sections. This typo does not exist in the official schema.xml that solr is using. -Original Message- From: Trevor Handley [mailto:hand...@civicplus.com] Sent: Monday, December 16, 2013 12:24 PM To: solr-user@lucene.apache.org Subject: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup Hello, I'm working with SolrCloud and trying to integrate the Suggester functionality http://wiki.apache.org/solr/Suggester. I've configured the requestHandler and searchComponent sections of solrconfig.xml, and added a new fieldtype and field to schema.xml. These documents are attached to this message. Background: This is a change I'm trying to make to a currently working/stable version of Solr 4.6 that has nearly 1.5 million unique documents in the index. The whole architecture is a single SolrCloud collection with 2 core shards that are replicated for a total of 4 cores. The shard1_slice1 core and shard2_slice2 core exist on one physical server, and shard1_slice2 core and shard2_slice1 core exist on a separate physical server. When I try to restart solr with suggester enabled then I get java error java.lang.ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup I tried a few different suggester classes but they all fail to load with the same message here. I verified that my .jar files do contain that class and are in the correct lib directory using a script that searches .jar files for a class name: [solr@Searchnode-001 ~]$ ./findclass.sh /opt/solrcloud/lib/ TSTLookup /opt/solrcloud/lib/solr-core-4.6.0.jar:org/apache/solr/spelling/suggest/tst/TSTLookupFactory.class /opt/solrcloud/lib/lucene-suggest-4.6-SNAPSHOT.jar:org/apache/lucene/search/suggest/tst/TSTLookup.class And here's a listing of the jar files in my lib directory: activation-1.1.jar AlchemyAPIAnnotator-2.3.1.jar apache-mime4j-core-0.7.2.jar apache-mime4j-dom-0.7.2.jar attributes-binder-1.2.0.jar bcmail-jdk15-1.45.jar bcprov-jdk15-1.45.jar boilerpipe-1.1.0.jar carrot2-mini-3.8.0.jar commons-beanutils-1.7.0.jar commons-collections-3.2.1.jar commons-compress-1.4.1.jar commons-digester-2.0.jar dom4j-1.6.1.jar fontbox-1.8.1.jar hppc-0.5.2.jar icu4j-49.1.jar isoparser-1.0-RC-1.jar jackson-core-asl-1.7.4.jar jackson-mapper-asl-1.7.4.jar jdom-1.0.jar jempbox-1.8.1.jar jetty-continuation-8.1.10.v20130312.jar jetty-deploy-8.1.10.v20130312.jar jetty-http-8.1.10.v20130312.jar jetty-io-8.1.10.v20130312.jar jetty-jmx-8.1.10.v20130312.jar jetty-security-8.1.10.v20130312.jar jetty-server-8.1.10.v20130312.jar jetty-servlet-8.1.10.v20130312.jar jetty-util-8.1.10.v20130312.jar jetty-webapp-8.1.10.v20130312.jar jetty-xml-8.1.10.v20130312.jar jsonic-1.2.7.jar juniversalchardet-1.0.3.jar langdetect-1.1-20120112.jar lucene-analyzers-common-4.6-SNAPSHOT.jar lucene-analyzers-kuromoji-4.6-SNAPSHOT.jar lucene-analyzers-phonetic-4.6-SNAPSHOT.jar lucene-codecs-4.6-SNAPSHOT.jar lucene-core-4.6-SNAPSHOT.jar lucene-grouping-4.6-SNAPSHOT.jar lucene-highlighter-4.6-SNAPSHOT.jar lucene-join-4.6-SNAPSHOT.jar lucene-memory-4.6-SNAPSHOT.jar lucene-misc-4.6-SNAPSHOT.jar lucene-queries-4.6-SNAPSHOT.jar lucene-queryparser-4.6-SNAPSHOT.jar lucene-spatial-4.6-SNAPSHOT.jar lucene-suggest-4.6-SNAPSHOT.jar mahout-collections-1.0.jar mahout-math-0.6.jar mail-1.4.1.jar metadata-extractor-2.6.2.jar morfologik-fsa-1.7.1.jar morfologik-polish-1.7.1.jar morfologik-stemming-1.7.1.jar netcdf-4.2-min.jar OpenCalaisAnnotator-2.3.1.jar pdfbox-1.8.1.jar poi-3.9.jar poi-ooxml-3.9.jar poi-ooxml-schemas-3.9.jar poi-scratchpad-3.9.jar rome-0.9.jar servlet-api-3.0.jar simple-xml-2.7.jar solr-analysis-extras-4.6.0.jar solr-cell-4.6.0.jar solr-clustering-4.6.0.jar solr-core-4.6.0.jar solr-dataimporthandler-4.6.0.jar solr-dataimporthandler-extras-4.6.0.jar solr-langid-4.6.0.jar solr-solrj-4.6.0.jar solr-test-framework-4.6.0.jar solr-uima-4.6.0.jar solr-velocity-4.6.0.jar Tagger-2.3.1.jar tagsoup-1.2.1.jar tika-core-1.4.jar tika-parsers-1.4.jar uimaj-core-2.3.1.jar velocity-1.7.jar velocity-tools-2.0.jar vorbis-java-core-0.1.jar vorbis-java-tika-0.1.jar WhitespaceTokenizer-2.3.1.jar xercesImpl-2.9.1.jar xmlbeans-2.3.0.jar xz-1.0.jar This is how I start solr with jetty: java -Dbootstrap_confdir=/opt/solrcloud/zkBootstrapConfigs/ -Dcollection.configName=CP_Search -DzkHost=zookeeper-001:2181,zookeeper-002:2181,zookeeper-003:2181 -Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.port=3000 -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.authenticate=false -server -Xms3g -Xmx6g -Xmn2g -XX:ParallelGCThreads=20 -XX:SurvivorRatio=8 -XX:TargetSurvivorRatio=90 -XX:MaxTenuringThreshold=31 -XX:+UseConcMarkSweepGC -XX:+UseParNewGC -XX:+CMSParallelRemarkEnabled -XX:+DisableExplicitGC -Djetty.port=8501 -Djetty.home=/opt/solrcloud/ -Dsolr.solr.home=/opt/solrcloud/ -DnumShards=2 -jar
PostingsSolrHighlighter
Hi @all, i am playing with the PostingsSolrHighlighter. I'm running solr 4.6.0 and my configuration is from here: https://lucene.apache.org/solr/4_6_0/solr-core/org/apache/solr/highlight/PostingsSolrHighlighter.html Search query and result (not working): http://pastebin.com/13Uan0ZF Schema (not complete): http://pastebin.com/JGa38UDT Search query and result (working): http://pastebin.com/4CP8XKnr Solr config: searchComponent class=solr.HighlightComponent name=highlight highlighting class=org.apache.solr.highlight.PostingsSolrHighlighter/ /searchComponent So this is working just fine, but now i have some questions: 1.) With the old default highlighter component it was possible to search in searchable_text and to retrive highlighted text. This is essential, because we use copyfield to put almost everything to searchable_text (title, subtitle, description, ...) 2.) I can't get ellipsis working i tried hl.tag.ellipsis=..., f.text.hl.tag.ellipsis=..., configuring it in RequestHandler noting seems to work, maxAnalyzedChars is just cutting the sentence? Kind Regards Josip Delic smime.p7s Description: S/MIME Cryptographic Signature
RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup
There have been some recent refactorings in this area of the code. The following class name should work: org.apache.solr.spelling.suggest.tst.TSTLookupFactory Cheers, Timothy Potter Sr. Software Engineer, LucidWorks www.lucidworks.com From: Trevor Handley hand...@civicplus.com Sent: Monday, December 16, 2013 11:27 AM To: solr-user@lucene.apache.org Subject: RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup Also, I'm aware that there's two typos in my schema.xml attached. I forgot to remove the linebreak \ character from the two splitOnCaseChange sections. This typo does not exist in the official schema.xml that solr is using. -Original Message- From: Trevor Handley [mailto:hand...@civicplus.com] Sent: Monday, December 16, 2013 12:24 PM To: solr-user@lucene.apache.org Subject: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup Hello, I'm working with SolrCloud and trying to integrate the Suggester functionality http://wiki.apache.org/solr/Suggester. I've configured the requestHandler and searchComponent sections of solrconfig.xml, and added a new fieldtype and field to schema.xml. These documents are attached to this message. Background: This is a change I'm trying to make to a currently working/stable version of Solr 4.6 that has nearly 1.5 million unique documents in the index. The whole architecture is a single SolrCloud collection with 2 core shards that are replicated for a total of 4 cores. The shard1_slice1 core and shard2_slice2 core exist on one physical server, and shard1_slice2 core and shard2_slice1 core exist on a separate physical server. When I try to restart solr with suggester enabled then I get java error java.lang.ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup I tried a few different suggester classes but they all fail to load with the same message here. I verified that my .jar files do contain that class and are in the correct lib directory using a script that searches .jar files for a class name: [solr@Searchnode-001 ~]$ ./findclass.sh /opt/solrcloud/lib/ TSTLookup /opt/solrcloud/lib/solr-core-4.6.0.jar:org/apache/solr/spelling/suggest/tst/TSTLookupFactory.class /opt/solrcloud/lib/lucene-suggest-4.6-SNAPSHOT.jar:org/apache/lucene/search/suggest/tst/TSTLookup.class And here's a listing of the jar files in my lib directory: activation-1.1.jar AlchemyAPIAnnotator-2.3.1.jar apache-mime4j-core-0.7.2.jar apache-mime4j-dom-0.7.2.jar attributes-binder-1.2.0.jar bcmail-jdk15-1.45.jar bcprov-jdk15-1.45.jar boilerpipe-1.1.0.jar carrot2-mini-3.8.0.jar commons-beanutils-1.7.0.jar commons-collections-3.2.1.jar commons-compress-1.4.1.jar commons-digester-2.0.jar dom4j-1.6.1.jar fontbox-1.8.1.jar hppc-0.5.2.jar icu4j-49.1.jar isoparser-1.0-RC-1.jar jackson-core-asl-1.7.4.jar jackson-mapper-asl-1.7.4.jar jdom-1.0.jar jempbox-1.8.1.jar jetty-continuation-8.1.10.v20130312.jar jetty-deploy-8.1.10.v20130312.jar jetty-http-8.1.10.v20130312.jar jetty-io-8.1.10.v20130312.jar jetty-jmx-8.1.10.v20130312.jar jetty-security-8.1.10.v20130312.jar jetty-server-8.1.10.v20130312.jar jetty-servlet-8.1.10.v20130312.jar jetty-util-8.1.10.v20130312.jar jetty-webapp-8.1.10.v20130312.jar jetty-xml-8.1.10.v20130312.jar jsonic-1.2.7.jar juniversalchardet-1.0.3.jar langdetect-1.1-20120112.jar lucene-analyzers-common-4.6-SNAPSHOT.jar lucene-analyzers-kuromoji-4.6-SNAPSHOT.jar lucene-analyzers-phonetic-4.6-SNAPSHOT.jar lucene-codecs-4.6-SNAPSHOT.jar lucene-core-4.6-SNAPSHOT.jar lucene-grouping-4.6-SNAPSHOT.jar lucene-highlighter-4.6-SNAPSHOT.jar lucene-join-4.6-SNAPSHOT.jar lucene-memory-4.6-SNAPSHOT.jar lucene-misc-4.6-SNAPSHOT.jar lucene-queries-4.6-SNAPSHOT.jar lucene-queryparser-4.6-SNAPSHOT.jar lucene-spatial-4.6-SNAPSHOT.jar lucene-suggest-4.6-SNAPSHOT.jar mahout-collections-1.0.jar mahout-math-0.6.jar mail-1.4.1.jar metadata-extractor-2.6.2.jar morfologik-fsa-1.7.1.jar morfologik-polish-1.7.1.jar morfologik-stemming-1.7.1.jar netcdf-4.2-min.jar OpenCalaisAnnotator-2.3.1.jar pdfbox-1.8.1.jar poi-3.9.jar poi-ooxml-3.9.jar poi-ooxml-schemas-3.9.jar poi-scratchpad-3.9.jar rome-0.9.jar servlet-api-3.0.jar simple-xml-2.7.jar solr-analysis-extras-4.6.0.jar solr-cell-4.6.0.jar solr-clustering-4.6.0.jar solr-core-4.6.0.jar solr-dataimporthandler-4.6.0.jar solr-dataimporthandler-extras-4.6.0.jar solr-langid-4.6.0.jar solr-solrj-4.6.0.jar solr-test-framework-4.6.0.jar solr-uima-4.6.0.jar solr-velocity-4.6.0.jar Tagger-2.3.1.jar tagsoup-1.2.1.jar tika-core-1.4.jar tika-parsers-1.4.jar uimaj-core-2.3.1.jar velocity-1.7.jar velocity-tools-2.0.jar vorbis-java-core-0.1.jar vorbis-java-tika-0.1.jar WhitespaceTokenizer-2.3.1.jar xercesImpl-2.9.1.jar xmlbeans-2.3.0.jar xz-1.0.jar This is how I start solr with jetty: java -Dbootstrap_confdir=/opt/solrcloud/zkBootstrapConfigs/ -Dcollection.configName=CP_Search
RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup
Brilliant, thanks Timothy! Changing the solrconfig.xml lookupImpl (not className) to the org.apache.solr.spelling.suggest.tst.TSTLookupFactory fixed this issue for me. Thanks, Trevor -Original Message- From: Tim Potter [mailto:tim.pot...@lucidworks.com] Sent: Monday, December 16, 2013 12:32 PM To: solr-user@lucene.apache.org Subject: RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup There have been some recent refactorings in this area of the code. The following class name should work: org.apache.solr.spelling.suggest.tst.TSTLookupFactory Cheers, Timothy Potter Sr. Software Engineer, LucidWorks www.lucidworks.com From: Trevor Handley hand...@civicplus.com Sent: Monday, December 16, 2013 11:27 AM To: solr-user@lucene.apache.org Subject: RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup Also, I'm aware that there's two typos in my schema.xml attached. I forgot to remove the linebreak \ character from the two splitOnCaseChange sections. This typo does not exist in the official schema.xml that solr is using. -Original Message- From: Trevor Handley [mailto:hand...@civicplus.com] Sent: Monday, December 16, 2013 12:24 PM To: solr-user@lucene.apache.org Subject: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup Hello, I'm working with SolrCloud and trying to integrate the Suggester functionality http://wiki.apache.org/solr/Suggester. I've configured the requestHandler and searchComponent sections of solrconfig.xml, and added a new fieldtype and field to schema.xml. These documents are attached to this message. Background: This is a change I'm trying to make to a currently working/stable version of Solr 4.6 that has nearly 1.5 million unique documents in the index. The whole architecture is a single SolrCloud collection with 2 core shards that are replicated for a total of 4 cores. The shard1_slice1 core and shard2_slice2 core exist on one physical server, and shard1_slice2 core and shard2_slice1 core exist on a separate physical server. When I try to restart solr with suggester enabled then I get java error java.lang.ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup I tried a few different suggester classes but they all fail to load with the same message here. I verified that my .jar files do contain that class and are in the correct lib directory using a script that searches .jar files for a class name: [solr@Searchnode-001 ~]$ ./findclass.sh /opt/solrcloud/lib/ TSTLookup /opt/solrcloud/lib/solr-core-4.6.0.jar:org/apache/solr/spelling/suggest/tst/TSTLookupFactory.class /opt/solrcloud/lib/lucene-suggest-4.6-SNAPSHOT.jar:org/apache/lucene/search/suggest/tst/TSTLookup.class And here's a listing of the jar files in my lib directory: activation-1.1.jar AlchemyAPIAnnotator-2.3.1.jar apache-mime4j-core-0.7.2.jar apache-mime4j-dom-0.7.2.jar attributes-binder-1.2.0.jar bcmail-jdk15-1.45.jar bcprov-jdk15-1.45.jar boilerpipe-1.1.0.jar carrot2-mini-3.8.0.jar commons-beanutils-1.7.0.jar commons-collections-3.2.1.jar commons-compress-1.4.1.jar commons-digester-2.0.jar dom4j-1.6.1.jar fontbox-1.8.1.jar hppc-0.5.2.jar icu4j-49.1.jar isoparser-1.0-RC-1.jar jackson-core-asl-1.7.4.jar jackson-mapper-asl-1.7.4.jar jdom-1.0.jar jempbox-1.8.1.jar jetty-continuation-8.1.10.v20130312.jar jetty-deploy-8.1.10.v20130312.jar jetty-http-8.1.10.v20130312.jar jetty-io-8.1.10.v20130312.jar jetty-jmx-8.1.10.v20130312.jar jetty-security-8.1.10.v20130312.jar jetty-server-8.1.10.v20130312.jar jetty-servlet-8.1.10.v20130312.jar jetty-util-8.1.10.v20130312.jar jetty-webapp-8.1.10.v20130312.jar jetty-xml-8.1.10.v20130312.jar jsonic-1.2.7.jar juniversalchardet-1.0.3.jar langdetect-1.1-20120112.jar lucene-analyzers-common-4.6-SNAPSHOT.jar lucene-analyzers-kuromoji-4.6-SNAPSHOT.jar lucene-analyzers-phonetic-4.6-SNAPSHOT.jar lucene-codecs-4.6-SNAPSHOT.jar lucene-core-4.6-SNAPSHOT.jar lucene-grouping-4.6-SNAPSHOT.jar lucene-highlighter-4.6-SNAPSHOT.jar lucene-join-4.6-SNAPSHOT.jar lucene-memory-4.6-SNAPSHOT.jar lucene-misc-4.6-SNAPSHOT.jar lucene-queries-4.6-SNAPSHOT.jar lucene-queryparser-4.6-SNAPSHOT.jar lucene-spatial-4.6-SNAPSHOT.jar lucene-suggest-4.6-SNAPSHOT.jar mahout-collections-1.0.jar mahout-math-0.6.jar mail-1.4.1.jar metadata-extractor-2.6.2.jar morfologik-fsa-1.7.1.jar morfologik-polish-1.7.1.jar morfologik-stemming-1.7.1.jar netcdf-4.2-min.jar OpenCalaisAnnotator-2.3.1.jar pdfbox-1.8.1.jar poi-3.9.jar poi-ooxml-3.9.jar poi-ooxml-schemas-3.9.jar poi-scratchpad-3.9.jar rome-0.9.jar servlet-api-3.0.jar simple-xml-2.7.jar solr-analysis-extras-4.6.0.jar solr-cell-4.6.0.jar solr-clustering-4.6.0.jar solr-core-4.6.0.jar solr-dataimporthandler-4.6.0.jar solr-dataimporthandler-extras-4.6.0.jar solr-langid-4.6.0.jar solr-solrj-4.6.0.jar solr-test-framework-4.6.0.jar solr-uima-4.6.0.jar solr-velocity-4.6.0.jar Tagger-2.3.1.jar
RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup
Awesome ... I'll update the Wiki to reflect the new class names. Timothy Potter Sr. Software Engineer, LucidWorks www.lucidworks.com From: Trevor Handley hand...@civicplus.com Sent: Monday, December 16, 2013 11:44 AM To: solr-user@lucene.apache.org Subject: RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup Brilliant, thanks Timothy! Changing the solrconfig.xml lookupImpl (not className) to the org.apache.solr.spelling.suggest.tst.TSTLookupFactory fixed this issue for me. Thanks, Trevor -Original Message- From: Tim Potter [mailto:tim.pot...@lucidworks.com] Sent: Monday, December 16, 2013 12:32 PM To: solr-user@lucene.apache.org Subject: RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup There have been some recent refactorings in this area of the code. The following class name should work: org.apache.solr.spelling.suggest.tst.TSTLookupFactory Cheers, Timothy Potter Sr. Software Engineer, LucidWorks www.lucidworks.com From: Trevor Handley hand...@civicplus.com Sent: Monday, December 16, 2013 11:27 AM To: solr-user@lucene.apache.org Subject: RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup Also, I'm aware that there's two typos in my schema.xml attached. I forgot to remove the linebreak \ character from the two splitOnCaseChange sections. This typo does not exist in the official schema.xml that solr is using. -Original Message- From: Trevor Handley [mailto:hand...@civicplus.com] Sent: Monday, December 16, 2013 12:24 PM To: solr-user@lucene.apache.org Subject: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup Hello, I'm working with SolrCloud and trying to integrate the Suggester functionality http://wiki.apache.org/solr/Suggester. I've configured the requestHandler and searchComponent sections of solrconfig.xml, and added a new fieldtype and field to schema.xml. These documents are attached to this message. Background: This is a change I'm trying to make to a currently working/stable version of Solr 4.6 that has nearly 1.5 million unique documents in the index. The whole architecture is a single SolrCloud collection with 2 core shards that are replicated for a total of 4 cores. The shard1_slice1 core and shard2_slice2 core exist on one physical server, and shard1_slice2 core and shard2_slice1 core exist on a separate physical server. When I try to restart solr with suggester enabled then I get java error java.lang.ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup I tried a few different suggester classes but they all fail to load with the same message here. I verified that my .jar files do contain that class and are in the correct lib directory using a script that searches .jar files for a class name: [solr@Searchnode-001 ~]$ ./findclass.sh /opt/solrcloud/lib/ TSTLookup /opt/solrcloud/lib/solr-core-4.6.0.jar:org/apache/solr/spelling/suggest/tst/TSTLookupFactory.class /opt/solrcloud/lib/lucene-suggest-4.6-SNAPSHOT.jar:org/apache/lucene/search/suggest/tst/TSTLookup.class And here's a listing of the jar files in my lib directory: activation-1.1.jar AlchemyAPIAnnotator-2.3.1.jar apache-mime4j-core-0.7.2.jar apache-mime4j-dom-0.7.2.jar attributes-binder-1.2.0.jar bcmail-jdk15-1.45.jar bcprov-jdk15-1.45.jar boilerpipe-1.1.0.jar carrot2-mini-3.8.0.jar commons-beanutils-1.7.0.jar commons-collections-3.2.1.jar commons-compress-1.4.1.jar commons-digester-2.0.jar dom4j-1.6.1.jar fontbox-1.8.1.jar hppc-0.5.2.jar icu4j-49.1.jar isoparser-1.0-RC-1.jar jackson-core-asl-1.7.4.jar jackson-mapper-asl-1.7.4.jar jdom-1.0.jar jempbox-1.8.1.jar jetty-continuation-8.1.10.v20130312.jar jetty-deploy-8.1.10.v20130312.jar jetty-http-8.1.10.v20130312.jar jetty-io-8.1.10.v20130312.jar jetty-jmx-8.1.10.v20130312.jar jetty-security-8.1.10.v20130312.jar jetty-server-8.1.10.v20130312.jar jetty-servlet-8.1.10.v20130312.jar jetty-util-8.1.10.v20130312.jar jetty-webapp-8.1.10.v20130312.jar jetty-xml-8.1.10.v20130312.jar jsonic-1.2.7.jar juniversalchardet-1.0.3.jar langdetect-1.1-20120112.jar lucene-analyzers-common-4.6-SNAPSHOT.jar lucene-analyzers-kuromoji-4.6-SNAPSHOT.jar lucene-analyzers-phonetic-4.6-SNAPSHOT.jar lucene-codecs-4.6-SNAPSHOT.jar lucene-core-4.6-SNAPSHOT.jar lucene-grouping-4.6-SNAPSHOT.jar lucene-highlighter-4.6-SNAPSHOT.jar lucene-join-4.6-SNAPSHOT.jar lucene-memory-4.6-SNAPSHOT.jar lucene-misc-4.6-SNAPSHOT.jar lucene-queries-4.6-SNAPSHOT.jar lucene-queryparser-4.6-SNAPSHOT.jar lucene-spatial-4.6-SNAPSHOT.jar lucene-suggest-4.6-SNAPSHOT.jar mahout-collections-1.0.jar mahout-math-0.6.jar mail-1.4.1.jar metadata-extractor-2.6.2.jar morfologik-fsa-1.7.1.jar morfologik-polish-1.7.1.jar morfologik-stemming-1.7.1.jar netcdf-4.2-min.jar OpenCalaisAnnotator-2.3.1.jar pdfbox-1.8.1.jar poi-3.9.jar poi-ooxml-3.9.jar
Re: OutOfMemoryError in RamUsageEstimator in solr 4.6
Hi Shawn, thanks for your reply. But we don't think that this is really a OOM error, because we already increased the heap to 64gb and the OOM occurs at a usage of 30-40gb. So solr would allocate more than 20gb at once. this sounds a little bit too much. Furthermore we found Lucene45DocValuesProducer and wondered why it exists and isn't used here. Lucene45DocValuesProducer.ramBytesUsed() also looks different to Lucene42DocValuesProducer.ramBytesUsed() Best Regards Torben 2013/12/16 Shawn Heisey s...@elyograg.org On 12/16/2013 2:34 AM, Torben Greulich wrote: we get a OutOfMemoryError in RamUsageEstimator and are a little bit confused about the error. We are using solr 4.6 and are confused about the Lucene42DocValuesProducer. We checked current solr code and found that Lucene42NormsFormat will be returned as NormFormat in Lucene46Codec and so the Lucene42DocValuesProducer will be returned. Is there any special reason not to have a higher version of NormsFormat? The format for that part of the index apparently hasn't changed since Lucene 4.2. Something else must have changed about the default index format, or I expect that there would not be a 4.6 specific Codec object. From what I've seen, the verion number encoded in the class name doesn't change until there's a new class needed for the default codec. A java OutOfMemoryError means that you need to increase the max heap when you start the program. It looks like you're trying to get a core status, so this is probably happening when using the admin UI. To put it quite simply, your Java max heap is not big enough to handle what Solr has been asked to do. http://wiki.apache.org/solr/SolrPerformanceProblems#Java_Heap Thanks, Shawn -- Torben Greulich · Entwickler Backend Development shopping24 internet group Wandsbeker Straße 3-7 · 22172 Hamburg Telefon: +49 (0) 40 6461 · Fax: +49 (0) 40 6461 7879 torben.greul...@s24.com · www.s24.com AG Hamburg HRB 63371 vertreten durch Dr. Björn Schäfers und Martin Mildner
Re: Possible parent/child query bug
On Fri, Nov 22, 2013 at 6:18 PM, Neil Ireson n.ire...@sheffield.ac.ukwrote: If the child of” query matches both parent and child docs it returns the child documents but a spurious numFound. follow up https://issues.apache.org/jira/browse/SOLR-5553 -- Sincerely yours Mikhail Khludnev Principal Engineer, Grid Dynamics http://www.griddynamics.com mkhlud...@griddynamics.com
Question about replication problem
I have a multi-core master/slave configuration that's showing unexpected replication behavior for one core; other cores are replicating without problems. The master is running Solr 4.1; one slave is running 4.1 under Tomcat, and another (for testing) is running 4.6 under Jetty. These are exhibiting similar problems. The core on the master is at generation 4; this is reported as both the generation for searching and replication using the replication API on the slave. When the slave attempts to replicate, it fails to load the file list required for replication, and logs the message: INFO: Master's generation: 4 INFO: Slave's generation: 5 INFO: Starting replication process SEVERE: No files to download for index generation: 4 (Timestamps removed for readability.) I have verified that the slave returns empty results where the master returns non-empty results; so this isn't a matter of reporting being incorrect. I'm not sure how to diagnose this; what can I check for to understand the actual problem? (I have Googled for relevant issues or mailing list traffic, but nothing actually explained what's happening or how to correct it.) -Fred -- Fred L. Drake, Jr.fdrake at gmail.com A storm broke loose in my mind. --Albert Einstein
Off-line search on mobile devices
Hello. We are planning to offer search as an embedded functionality into mobile/low-power devices. The main requirement are: - ability to index and search documents available on the mobile device, - no need of internet access, - lightweight, low footprint and fast We are looking into various options. As I understand it, Solr would be way too heavy for mobile devices. Has anyone used Lucene/Solr for off-line search on mobile devices? Are there better alternatives for off-line full-text search? Many thanks. Arcadius.
Re: Off-line search on mobile devices
1. Which platform are you looking at? Android, iOS, other? If you are on Android, you can directly use lucene to build an embedded solution for search. Depending upon your need, that can offer a small enough footprint. We've done some work around embedding lucene for a specific application on Android, happy to brainstorm offline. Thanks, Sameer. -- http://measuredsearch.com On Mon, Dec 16, 2013 at 3:07 PM, Arcadius Ahouansou arcad...@menelic.comwrote: Hello. We are planning to offer search as an embedded functionality into mobile/low-power devices. The main requirement are: - ability to index and search documents available on the mobile device, - no need of internet access, - lightweight, low footprint and fast We are looking into various options. As I understand it, Solr would be way too heavy for mobile devices. Has anyone used Lucene/Solr for off-line search on mobile devices? Are there better alternatives for off-line full-text search? Many thanks. Arcadius. -- Sameer Maggon Founder, Measured Search m: 310.344.7266 tw: @measuredsearch w: http://www.measuredsearch.com
how to remove documents which have been deleted from the database
Hi, Glassfish 3.1.2.2 Solr 4.5 Zookeeper 3.4.5 We have set up a SolrCloud with 4 Solr nodes and 3 zookeeper instances. I have 5 cores with 1 shard/4 replica setup on each of them. One of our core is very small, and it takes less than one minute to index. We run full import on it every hour with commit= true and clear =false but full import does not delete documents which are no more there in database. if I run the import with clear =true it will delete the index first and shows 0 records during the import is in progress. For delta we will have to create a new table in database which will keep track of deleted records. we dont want to go that route. Is there a way, where, we can run a full import with clear= true and solr only update the data after the import is completed. Regards, Kaustubh -- View this message in context: http://lucene.472066.n3.nabble.com/how-to-remove-documents-which-have-been-deleted-from-the-database-tp4107015.html Sent from the Solr - User mailing list archive at Nabble.com.
zookeeper timeout issue
Hi, Following warning message is filling our application logs very rapidly. This statement is printed every time application talks with zookeeper. [#|2013-12-13T08:33:03.023-0800|INFO|glassfish3.1.2|javax.enterprise.system.std.com.sun.enterprise.server.logging|_ThreadID=64;_ThreadName=Thread-2;|2013-12-13 08:33:03,023 WARN org.apache.zookeeper.ClientCnxn - Session 0x342e9013f2d0063 for server null, unexpected error, closing socket connection and attempting reconnect java.lang.NullPointerException: key can't be null at java.lang.System.checkKey(System.java:771) at java.lang.System.getProperty(System.java:647) at org.apache.zookeeper.client.ZooKeeperSaslClient.init(ZooKeeperSaslClient.java:133) at org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:943) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993) I am getting following INFO on zookeeper logs [myid:1] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:ZooKeeperServer@812] - Refusing session request for client /IP1:15703 as it has seen zxid 0x1c1103 our last zxid is 0x40348 client must try another server 2013-12-16 15:45:56,999 [myid:1] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@1001] - Closed socket connection for client /10.163.160.78:15703 (no session established for client) it could be related to this problem https://issues.apache.org/jira/browse/ZOOKEEPER-1237 or http://mail-archives.apache.org/mod_mbox/zookeeper-user/201208.mbox/%3CCANLc_9Jwieyyig=yg1yvczaeobc8swwj3fqd4x993ryrpod...@mail.gmail.com%3E Is there a solution to this problem or I will have to wait for next zookeeper release. Regards, Kaustubh -- View this message in context: http://lucene.472066.n3.nabble.com/zookeeper-timeout-issue-tp4107016.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: how to remove documents which have been deleted from the database
If your index is that small then use clean=true and let DataImportHandler clear all documents at the start of the import. Don't bother with delta tables for such a small index. On Tue, Dec 17, 2013 at 6:10 AM, kaustubh147 kaustubh.j...@gmail.com wrote: Hi, Glassfish 3.1.2.2 Solr 4.5 Zookeeper 3.4.5 We have set up a SolrCloud with 4 Solr nodes and 3 zookeeper instances. I have 5 cores with 1 shard/4 replica setup on each of them. One of our core is very small, and it takes less than one minute to index. We run full import on it every hour with commit= true and clear =false but full import does not delete documents which are no more there in database. if I run the import with clear =true it will delete the index first and shows 0 records during the import is in progress. For delta we will have to create a new table in database which will keep track of deleted records. we dont want to go that route. Is there a way, where, we can run a full import with clear= true and solr only update the data after the import is completed. Regards, Kaustubh -- View this message in context: http://lucene.472066.n3.nabble.com/how-to-remove-documents-which-have-been-deleted-from-the-database-tp4107015.html Sent from the Solr - User mailing list archive at Nabble.com. -- Regards, Shalin Shekhar Mangar.
Re: Storing MYSQL DATETIME field in solr as String
I'm sorry. I thought you wanted to parse a date stored as string into a java.util.Date. Clearly, you are trying to go the other way round. There's nothing in DIH which will convert a mysql date to a string in specific format. You will need to write a custom transformer either in javascript or in java to do this. http://wiki.apache.org/solr/DIHCustomTransformer http://wiki.apache.org/solr/DataImportHandler#ScriptTransformer On Mon, Dec 16, 2013 at 1:34 PM, manju16832003 manju16832...@gmail.com wrote: OK here are the scenarios I tried. *Scenario - 1: * dih.xml (aka data-config.xml) entity dataSource=solr name=listing query=... transformer=DateFormatTransformer field name=quot;publish_datequot; column=quot;publish_datequot; xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd'T'HH:mm:ss* locale=en/ schema.xml field name=publish_date type=date indexed=true stored=true multiValued=false default=NOW / It does not work. Throws an following exception WARN org.apache.solr.handler.dataimport.DateFormatTransformer – Could not parse a Date field java.text.ParseException: Unparseable date: 2013-12-05 15:40:03.0 *Scenario - 2* dih.xml (aka data-config.xml) entity dataSource=solr name=listing query=... transformer=DateFormatTransformer field name=quot;publish_datequot; column=quot;publish_datequot; xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd HH:mm:ss* locale=en/ schema.xml field name=publish_date type=date indexed=true stored=true multiValued=false default=NOW / Output is date name=publish_date2012-12-05T07:38:27Z/date *Scenario - 3* dih.xml (aka data-config.xml) entity dataSource=solr name=listing query=... transformer=DateFormatTransformer field name=quot;publish_datequot; column=quot;publish_datequot; xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd HH:mm:ss* locale=en/ schema.xml field name=publish_date type=string indexed=true stored=true multiValued=false default=NOW / Output is str name=publish_dateThu Dec 05 15:38:27 MYT 2013/str My expectation was *2012-12-05 07:38:27* (MySQL DateTime format) -- View this message in context: http://lucene.472066.n3.nabble.com/Storing-MYSQL-DATETIME-field-in-solr-as-String-tp4106836p4106854.html Sent from the Solr - User mailing list archive at Nabble.com. -- Regards, Shalin Shekhar Mangar.
Cannot create core
Hi I get an weird problem. I try to create a core within Solr4.6. Firstly, on my solr web server tomcat, a log come out[1]: Then lots of Overseer Info logs come out as [2]. And then the creation failed. I have also notice that there is a lot of qn in the Overseer on the zookeeper: [zk: localhost:2181(CONNECTED) 45] ls /overseer/queue [qn-38, qn-37, qn-39, qn-34, qn-33, qn-36, qn-44, qn-35, qn-42, qn-43, qn-30, qn-31, qn-40, qn-32, qn-41] Why are there so many qn.? I do not know the reason to the core creation failed. Please help . [1] .. 293699 [http-bio-8081-exec-1] INFO org.apache.solr.cloud.ZkController – publishing core=reportCore_201310 state=down 293700 [http-bio-8081-exec-1] INFO org.apache.solr.cloud.ZkController – numShards not found on descriptor - reading it from system property 293704 [http-bio-8081-exec-1] INFO org.apache.solr.cloud.ZkController – waiting to find shard id in clusterstate for reportCore_201310 [2]--- 136661 [Thread-2] INFO org.apache.solr.common.cloud.ZkStateReader – Updating cloud state from ZooKeeper... 136663 [Thread-2] INFO org.apache.solr.cloud.Overseer – Update state numShards=null message={ operation:state, state:down, base_url:http://10.1.22.1:8080/solr;, core:reportCore_201310, roles:null, node_name:10.1.22.1:8080_solr, shard:reportCore_201310, shard_range:null, shard_state:active, shard_parent:null, collection:repCore, numShards:null, core_node_name:reportCore_201310} 136663 [Thread-2] ERROR org.apache.solr.cloud.Overseer – Exception in Overseer main queue loop java.lang.NullPointerException at org.apache.solr.cloud.Overseer$ClusterStateUpdater.updateState(Overseer.java:360) at org.apache.solr.cloud.Overseer$ClusterStateUpdater.processMessage(Overseer.java:204) at org.apache.solr.cloud.Overseer$ClusterStateUpdater.run(Overseer.java:166) at java.lang.Thread.run(Thread.java:662)
Solr cores across multiple machines
Hi, In my project, we are doing full index on dedicated machine and the index will be copied to other search serving machine. For this, we are copying the data folder from indexing machine to serving machine manually. Now, we wanted to use Solr's SWAP configuration to do this job. Looks like the SWAP will work between the cores. Based on our setup, any one has any idea how to move the data from indexing machine to serving machine? Is there any other alternatives? Regards, Siva -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-cores-across-multiple-machines-tp4107035.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Storing MYSQL DATETIME field in solr as String
To me, the obvious way of doing this would be to CAST the DATETIME to CHAR(n), or (probably better) use DATE_FORMAT(). On Tue, Dec 17, 2013 at 5:21 AM, Shalin Shekhar Mangar shalinman...@gmail.com wrote: I'm sorry. I thought you wanted to parse a date stored as string into a java.util.Date. Clearly, you are trying to go the other way round. There's nothing in DIH which will convert a mysql date to a string in specific format. You will need to write a custom transformer either in javascript or in java to do this. http://wiki.apache.org/solr/DIHCustomTransformer http://wiki.apache.org/solr/DataImportHandler#ScriptTransformer On Mon, Dec 16, 2013 at 1:34 PM, manju16832003 manju16832...@gmail.com wrote: OK here are the scenarios I tried. *Scenario - 1: * dih.xml (aka data-config.xml) entity dataSource=solr name=listing query=... transformer=DateFormatTransformer field name=quot;publish_datequot; column=quot;publish_datequot; xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd'T'HH:mm:ss* locale=en/ schema.xml field name=publish_date type=date indexed=true stored=true multiValued=false default=NOW / It does not work. Throws an following exception WARN org.apache.solr.handler.dataimport.DateFormatTransformer – Could not parse a Date field java.text.ParseException: Unparseable date: 2013-12-05 15:40:03.0 *Scenario - 2* dih.xml (aka data-config.xml) entity dataSource=solr name=listing query=... transformer=DateFormatTransformer field name=quot;publish_datequot; column=quot;publish_datequot; xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd HH:mm:ss* locale=en/ schema.xml field name=publish_date type=date indexed=true stored=true multiValued=false default=NOW / Output is date name=publish_date2012-12-05T07:38:27Z/date *Scenario - 3* dih.xml (aka data-config.xml) entity dataSource=solr name=listing query=... transformer=DateFormatTransformer field name=quot;publish_datequot; column=quot;publish_datequot; xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd HH:mm:ss* locale=en/ schema.xml field name=publish_date type=string indexed=true stored=true multiValued=false default=NOW / Output is str name=publish_dateThu Dec 05 15:38:27 MYT 2013/str My expectation was *2012-12-05 07:38:27* (MySQL DateTime format) -- View this message in context: http://lucene.472066.n3.nabble.com/Storing-MYSQL-DATETIME-field-in-solr-as-String-tp4106836p4106854.html Sent from the Solr - User mailing list archive at Nabble.com. -- Regards, Shalin Shekhar Mangar.
Re: External File Field
Hi http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,locality,property_price_min,field(adjlocality)q=*:*http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,locality,property_price_min,norm:pricenorm(1000,10,property_price_median),field(adjlocality)q=*:* I did it, it is giving me a message underfined field. I have placed my files at F:\solr\example\solr\collection1\data where the Solr data is maintained. Is there a way to check whether the files are read or not? Thanks Regards Mukund On Mon, Dec 16, 2013 at 8:48 PM, Ahmet Arslan iori...@yahoo.com wrote: Hi, You can request the value of the that field via fl=*,field(adjlocality) See more about it : https://cwiki.apache.org/confluence/display/solr/Working+with+External+Files+and+Processes Actually you can search it too with frange query parser. {!frange l=0 u=0}field(adjlocality) On Monday, December 16, 2013 5:05 PM, Mukundaraman valakumaresan muk...@8kmiles.com wrote: Hi Is there a way to find ,whether the External File Fields mentioned in the schema.xml is being used or whether Solr reads the value of those external fields. I am not sure how to use an External field, can I request the value of an External File Field in the field list or can I use in my custom function as shown below: http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,adjlocality http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,property_price_min,lanorm(%22%22,luxury_amenities_c,10),adjlocality ,custfunc(adjlocality). I didn't get any output for the external fields in the queries that I provide. Thanks Regards Mukund