Re: How to delete documents from a SOLR cloud / balance the shards in the cloud?
Stephan and all, I am evaluating this like you are. You may want to check http://www.tomkleinpeter.com/2008/03/17/programmers-toolbox-part-3-consistent-hashing/. I would appreciate if others can shed some light on this, too. Bests, James On Fri, Sep 10, 2010 at 6:07 AM, Stephan Raemy stephan.ra...@gmail.comwrote: Hi solr-cloud users, I'm currently setting up a solr-cloud/zookeeper instance and so far, everything works out fine. I downloaded the source from the cloud branch yesterday and build it from source. I've got 10 shards distributed across 4 servers and a zookeeper instance. Searching documents with the flag distrib=true works out and it returns the expected result. But here comes the tricky question. I will add new documents every day and therefore, I'd like to balance my shards to keep the system speedy. The Wiki says that one can calculate the hash of a document id and then determine the corresponding shard. But IMHO, this does not take into account that the cloud may become bigger or shrink over time by adding or removing shards. Obviously adding has a higher priority since one wants to reduce the shard size to improve the response time of distributed searches. When reading through the Wikis and existing documentation, it is still unclear to me how to do the following operations: - Modify/Delete a document stored in the cloud without having to store the document:shard mapping information outside of the cloud. I would expect something like shard attribute on each doc in the SOLR query result (activated/deactivated by a flag), so that i can query the SOLR cloud for a doc and then delete it on the specific shard. - Balance a cloud when adding/removing new shards or just balance them after many deletions. Of course there are solutions to this, but at the end, I'd love to have a true cloud where i do not have to worry about shard performance optimization. Hints are greatly appreciated. Cheers, Stephan
how to stress test solr
before stressing test, Should i close SolrCache? which tool u use? How to do stress test correctly? Any pointers? -- regards j.L ( I live in Shanghai, China)
Re: Is it problem? I use solr to search and index is made by lucene. (not EmbeddedSolrServer(wiki is old))
solr have much fieldtype, like: integer,long, double, sint, sfloat, tint,tfloat,,and more. but lucene not fieldtype,,just name and value, value only string. so i not sure is it a problem when i use solr to search( index made by lucene). -- regards j.L ( I live in Shanghai, China)
IndexMerge not found
i try http://wiki.apache.org/solr/MergingSolrIndexes system: win2003, jdk 1.6 Error information: Caused by: java.lang.ClassNotFoundException: org.apache.lucene.misc.IndexMergeTo ol at java.net.URLClassLoader$1.run(Unknown Source) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) at sun.misc.Launcher$AppClassLoader.loadClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) at java.lang.ClassLoader.loadClassInternal(Unknown Source) Could not find the main class: org/apache/lucene/misc/IndexMergeTool. Program w ill exit. -- regards j.L ( I live in Shanghai, China)
Re: IndexMerge not found
i use lucene-core-2.9-dev.jar, lucene-misc-2.9-dev.jar On Thu, Jul 2, 2009 at 2:02 PM, James liu liuping.ja...@gmail.com wrote: i try http://wiki.apache.org/solr/MergingSolrIndexes system: win2003, jdk 1.6 Error information: Caused by: java.lang.ClassNotFoundException: org.apache.lucene.misc.IndexMergeTo ol at java.net.URLClassLoader$1.run(Unknown Source) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) at sun.misc.Launcher$AppClassLoader.loadClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) at java.lang.ClassLoader.loadClassInternal(Unknown Source) Could not find the main class: org/apache/lucene/misc/IndexMergeTool. Program w ill exit. -- regards j.L ( I live in Shanghai, China) -- regards j.L ( I live in Shanghai, China)
Is it problem? I use solr to search and index is made by lucene. (not EmbeddedSolrServer(wiki is old))
I use solr to search and index is made by lucene. (not EmbeddedSolrServer(wiki is old)) Is it problem when i use solr to search? which the difference between Index(made by lucene and solr)? thks -- regards j.L ( I live in Shanghai, China)
does solr support summary
if user use keyword to search and get summary(auto generated by keyword)...like this doc filed: id, text id: 001 text: Open source is a development method for software that harnesses the power of distributed peer review and transparency of process. The promise of open source is better quality, higher reliability, more flexibility, lower cost, and an end to predatory vendor lock-in. if keyword is source,,summary is: Open source is a development...The promise of open source is better quality if keyword is power ,,,summary is: Open...harnesses the power of distributed peer review and transparency of process... just like google search results... and any advice will be appreciated. -- regards j.L ( I live in Shanghai, China)
Re: timeouts
*Collins: *i don't know what u wanna say? -- regards j.L ( I live in Shanghai, China)
Re: indexing Chienese langage
first: u not have to restart solr,,,u can use new data to replace old data and call solr to use new search..u can find something in shell script which with solr two: u not have to restart solr,,,just keep id is same..example: old id:1,title:hi, new id:1,title:welcome,,just index new data,,it will delete old data and insert new doc,,,like replace,,but it will use more time and resouce. u can find indexed doc number from solr admin page. On Fri, Jun 5, 2009 at 7:42 AM, Fer-Bj fernando.b...@gmail.com wrote: What we usually do to reindex is: 1. stop solr 2. rmdir -r data (that is to remove everything in /opt/solr/data/ 3. mkdir data 4. start solr 5. start reindex. with this we're sure about not having old copies or index.. To check the index size we do: cd data du -sh Otis Gospodnetic wrote: I can't tell what that analyzer does, but I'm guessing it uses n-grams? Maybe consider trying https://issues.apache.org/jira/browse/LUCENE-1629 instead? Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch - Original Message From: Fer-Bj fernando.b...@gmail.com To: solr-user@lucene.apache.org Sent: Thursday, June 4, 2009 2:20:03 AM Subject: Re: indexing Chienese langage We are trying SOLR 1.3 with Paoding Chinese Analyzer , and after reindexing the index size went from 1.5 Gb to 2.7 Gb. Is that some expected behavior ? Is there any switch or trick to avoid having a double + index file size? Koji Sekiguchi-2 wrote: CharFilter can normalize (convert) traditional chinese to simplified chinese or vice versa, if you define mapping.txt. Here is the sample of Chinese character normalization: https://issues.apache.org/jira/secure/attachment/12392639/character-normalization.JPG See SOLR-822 for the detail: https://issues.apache.org/jira/browse/SOLR-822 Koji revathy arun wrote: Hi, When I index chinese content using chinese tokenizer and analyzer in solr 1.3 ,some of the chinese text files are getting indexed but others are not. Since chinese has got many different language subtypes as in standard chinese,simplified chinese etc which of these does the chinese tokenizer support and is there any method to find the type of chiense language from the file? Rgds -- View this message in context: http://www.nabble.com/indexing-Chienese-langage-tp22033302p23864358.html Sent from the Solr - User mailing list archive at Nabble.com. -- View this message in context: http://www.nabble.com/indexing-Chienese-langage-tp22033302p23879730.html Sent from the Solr - User mailing list archive at Nabble.com. -- regards j.L ( I live in Shanghai, China)
Re: indexing Chienese langage
On Mon, Feb 16, 2009 at 4:30 PM, revathy arun revas...@gmail.com wrote: Hi, When I index chinese content using chinese tokenizer and analyzer in solr 1.3 ,some of the chinese text files are getting indexed but others are not. are u sure ur analyzer can do it good? if not sure, u can use analzyer link in solr admin page to check it Since chinese has got many different language subtypes as in standard chinese,simplified chinese etc which of these does the chinese tokenizer support and is there any method to find the type of chiense language from the file? Rgds -- regards j.L ( I live in Shanghai, China)
Re: Using Chinese / How to ?
1: modify ur schema.xml: like fieldtype name=text_cn class=solr.TextField analyzer class=chineseAnalyzer/ analyzer 2: add your field: field name=urfield type=text_cn indexd=true stored=true/ 3: add your analyzer to {solr_dir}\lib\ 4: rebuild newsolr and u will find it in {solr_dir}\dist 5: follow tutorial to setup solr 6: open your browser to solr admin page, find analyzer to check analyzer, it will tell u how to analyzer world, use which analyzer -- regards j.L ( I live in Shanghai, China)
Re: Using Chinese / How to ?
u means how to config solr which support chinese? Update problem? On Tuesday, June 2, 2009, Fer-Bj fernando.b...@gmail.com wrote: I'm sending 3 files: - schema.xml - solrconfig.xml - error.txt (with the error description) I can confirm by now that this error is due to invalid characters for the XML format (ASCII 0 or 11). However, this problem now is taking a different direction: how to start using the CJK instead of the english! http://www.nabble.com/file/p23825881/error.txt error.txt http://www.nabble.com/file/p23825881/solrconfig.xml solrconfig.xml http://www.nabble.com/file/p23825881/schema.xml schema.xml Grant Ingersoll-6 wrote: Can you provide details on the errors? I don't think we have a specific how to, but I wouldn't think it would be much different from 1.2 -Grant On May 31, 2009, at 10:31 PM, Fer-Bj wrote: Hello, is there any how to already created to get me up using SOLR 1.3 running for a chinese based website? Currently our site is using SOLR 1.2, and we tried to move into 1.3 but we couldn't complete our reindex as it seems like 1.3 is more strict when it comes to special chars. I would appreciate any help anyone may provide on this. Thanks!! -- View this message in context: http://www.nabble.com/Using-Chinese---How-to---tp23810129p23810129.html Sent from the Solr - User mailing list archive at Nabble.com. -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search -- View this message in context: http://www.nabble.com/Using-Chinese---How-to---tp23810129p23825881.html Sent from the Solr - User mailing list archive at Nabble.com. -- regards j.L ( I live in Shanghai, China)
Re: Solr multiple keyword search as google
U can find answer in tutorial or example On Tuesday, June 2, 2009, The Spider maheshmura...@rediffmail.com wrote: Hi, I am using solr nightly bind for my search. I have to search in the location field of the table which is not my default search field. I will briefly explain my requirement below: I want to get the same/similar result when I give location multiple keywords, say San jose ca USA or USA ca san jose or CA San jose USA (like that of google search). That means even if I rearranged the keywords of location I want to get proper results. Is there any way to do that? Thanks in advance -- View this message in context: http://www.nabble.com/Solr-multiple-keyword-search-as-google-tp23826278p23826278.html Sent from the Solr - User mailing list archive at Nabble.com. -- regards j.L ( I live in Shanghai, China)
Re: Query for Distributed search -
Up to your solr client. On Mon, Nov 24, 2008 at 1:24 PM, souravm [EMAIL PROTECTED] wrote: Hi, Looking for some insight on distributed search. Say I have an index distributed in 3 boxes and the index contains time and text data (typical log file). Each box has index for different timeline - say Box 1 for all Jan to April, Box 2 for May to August and Box 3 for Sep to Dec. Now if I try to search for a text string, will the search would happen in parallel in all 3 boxes or sequentially? Regards, Sourav CAUTION - Disclaimer * This e-mail contains PRIVILEGED AND CONFIDENTIAL INFORMATION intended solely for the use of the addressee(s). If you are not the intended recipient, please notify the sender by e-mail and delete the original message. Further, you are not to copy, disclose, or distribute this e-mail or its contents to any other person and any such actions are unlawful. This e-mail may contain viruses. Infosys has taken every reasonable precaution to minimize this risk, but is not liable for any damage you may sustain as a result of any virus in this e-mail. You should carry out your own virus checks before opening the e-mail or attachment. Infosys reserves the right to monitor and review the content of all messages sent to or from this e-mail address. Messages sent to or from this e-mail address may be stored on the Infosys e-mail system. ***INFOSYS End of Disclaimer INFOSYS*** -- regards j.L
Re: Newbe! Trying to run solr-1.3.0 under tomcat. Please help
check procedure: 1: rm -r $tomcat/webapps/* 2: rm -r $solr/data ,,,ur index data directory 3: check xml(any xml u modified) 4: start tomcat i had same error, but i forgot how to fix...so u can use my check procedure, i think it will help you i use tomcat+solr in win2003, freebsd, mac osx 10.5.5, they all work well -- regards j.L
Re: posting error in solr
first u sure the xml is utf-8,,and field value is utf-8,, second u should post xml by utf-8 my advice : All encoding use utf-8... it make my solr work well,,, i use chinese -- regards j.L
Re: sole 1.3: bug in phps response writer
i find url not same as the others -- regards j.L
Re: Illegal character in xml file
first, u should escape some string like (code by php) function escapeChars($string) { $string = str_replace(, amp;, $string); $string = str_replace(, lt;, $string); $string = str_replace(, gt;, $string); $string = str_replace(', apos;, $string); $string = str_replace('', quot;, $string); return $string; } second, u get xml(encode by utf-8) third, post it by utf-8(head string Content-Type: text/xml;charset=utf-8) if u don't know how to, maybe u can check solr client(u can find it in solr's wiki) Good Luck~ On Fri, Sep 19, 2008 at 4:33 PM, 李学健 [EMAIL PROTECTED] wrote: hi, all when i post xml files to solr, it's interrupted by this: Illegal character how can i deal with it ? is there any solution to ignore Illegal character in documents feeded ? thanks -- regards j.L
Re: CJK Analyzers for Solr
it seems good. On Dec 3, 2007 1:01 AM, Ken Krugler [EMAIL PROTECTED] wrote: Wunder - are you aware of any free dictionaries for either C or J or K? When I dealt with this in the past, I looked for something free, but found only commercial dictionaries. I would use data files from: http://ftp.monash.edu.au/pub/nihongo/00INDEX.html -- Ken Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch - Original Message From: Walter Underwood [EMAIL PROTECTED] To: solr-user@lucene.apache.org Sent: Wednesday, November 28, 2007 5:43:32 PM Subject: Re: CJK Analyzers for Solr With Ultraseek, we switched to a dictionary-based segmenter for Chinese because the N-gram highlighting wasn't acceptable to our Chinese customers. I guess it is something to check for each application. wunder On 11/27/07 10:46 PM, Otis Gospodnetic [EMAIL PROTECTED] wrote: For what it's worth I worked on indexing and searching a *massive* pile of data, a good portion of which was in CJ and some K. The n-gram approach was used for all 3 languages and the quality of search results, including highlighting was evaluated and okay-ed by native speakers of these languages. Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch - Original Message From: Walter Underwood [EMAIL PROTECTED] To: solr-user@lucene.apache.org Sent: Tuesday, November 27, 2007 2:41:38 PM Subject: Re: CJK Analyzers for Solr Dictionaries are surprisingly expensive to build and maintain and bi-gram is surprisingly effective for Chinese. See this paper: http://citeseer.ist.psu.edu/kwok97comparing.html I expect that n-gram indexing would be less effective for Japanese because it is an inflected language. Korean is even harder. It might work to break Korean into the phonetic subparts and use n-gram on those. You should not do term highlighting with any of the n-gram methods. The relevance can be very good, but the highlighting just looks dumb. wunder On 11/27/07 8:54 AM, Eswar K [EMAIL PROTECTED] wrote: Is there any specific reason why the CJK analyzers in Solr were chosen to be n-gram based instead of it being a morphological analyzer which is kind of implemented in Google as it considered to be more effective than the n-gram ones? Regards, Eswar On Nov 27, 2007 7:57 AM, Eswar K [EMAIL PROTECTED] wrote: thanks james... How much time does it take to index 18m docs? - EswarOn Nov 27, 2007 7:43 AM, James liu [EMAIL PROTECTED] wrote: i not use HYLANDA analyzer. i use je-analyzer and indexing at least 18m docs. i m sorry i only use chinese analyzer.On Nov 27, 2007 10:01 AM, Eswar K [EMAIL PROTECTED] wrote: What is the performance of these CJK analyzers (one in lucene and hylanda )? We would potentially be indexing millions of documents. James, We would have a look at hylanda too. What abt japanese and korean analyzers, any recommendations? - Eswar On Nov 27, 2007 7:21 AM, James liu [EMAIL PROTECTED] wrote: I don't think NGram is good method for Chinese. CJKAnalyzer of Lucene is 2-Gram. Eswar K: if it is chinese analyzer,,i recommend hylandaÅiwww.hylanda.comÅj,,,it is the best chinese analyzer and it not free. if u wanna free chinese analyzer, maybe u can try je-analyzer. it have some problem when using it. On Nov 27, 2007 5:56 AM, Otis Gospodnetic [EMAIL PROTECTED] wrote: Eswar, We've uses the NGram stuff that exists in Lucene's contrib/analyzers instead of CJK. Doesn't that allow you to do everything that the Chinese and CJK analyzers do? It's been a few months since I've looked at Chinese and CJK Analzyers, so I could be off. Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch - Original Message From: Eswar K [EMAIL PROTECTED] To: solr-user@lucene.apache.org Sent: Monday, November 26, 2007 8:30:52 AM Subject: CJK Analyzers for Solr Hi, Does Solr come with Language analyzers for CJK? If not, can you please direct me to some good CJK analyzers? Regards, Eswar -- regards jl -- regards jl -- Ken Krugler Krugle, Inc. +1 530-210-6378 If you can't find it, you can't fix it -- regards jl
Re: CJK Analyzers for Solr
I don't think NGram is good method for Chinese. CJKAnalyzer of Lucene is 2-Gram. Eswar K: if it is chinese analyzer,,i recommend hylanda(www.hylanda.com),,,it is the best chinese analyzer and it not free. if u wanna free chinese analyzer, maybe u can try je-analyzer. it have some problem when using it. On Nov 27, 2007 5:56 AM, Otis Gospodnetic [EMAIL PROTECTED] wrote: Eswar, We've uses the NGram stuff that exists in Lucene's contrib/analyzers instead of CJK. Doesn't that allow you to do everything that the Chinese and CJK analyzers do? It's been a few months since I've looked at Chinese and CJK Analzyers, so I could be off. Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch - Original Message From: Eswar K [EMAIL PROTECTED] To: solr-user@lucene.apache.org Sent: Monday, November 26, 2007 8:30:52 AM Subject: CJK Analyzers for Solr Hi, Does Solr come with Language analyzers for CJK? If not, can you please direct me to some good CJK analyzers? Regards, Eswar -- regards jl
Re: CJK Analyzers for Solr
if ur analyzer is standard, u can try use tokenize.(u can find the answer from analyzer source code and schema.xml) On Nov 27, 2007 9:39 AM, zx zhang [EMAIL PROTECTED] wrote: lance, The following is a instance schema fieldtype using solr1.2 and CJK package. And it works. As you said, CJK does parse cjk string in a bi-gram way, just like turning 'C1C2C3C4' into 'C1C2 C2C3 C3C4'. More to the point, it is worthwhile to mention that the index expand beyond tolerance to use cjk package, and it will take a long time to index document. For most enterprise applications, I think, it need a more effective string parser. fieldtype name=text_cjk class=solr.TextField analyzer class=org.apache.lucene.analysis.cjk.CJKAnalyzer/ /fieldtype On 11/27/07, Norskog, Lance [EMAIL PROTECTED] wrote: I notice this is in the future tense. Is the CJKTokenizer available yet? From what I can see, the CJK code should be a Filter instead anyway. Also, the ChineseFilter and CJKTokenizer do two different things. CJKTokenizer turns C1C2C3C4 into 'C1C2 C2C3 C3C4'. ChineseFilter (from 2001) turns C1C2 into 'C1 C2'. I hope someone who speaks Mandarin or Cantonese understands what this should do. Lance -Original Message- From: Eswar K [mailto:[EMAIL PROTECTED] Sent: Monday, November 26, 2007 10:28 AM To: solr-user@lucene.apache.org Subject: Re: CJK Analyzers for Solr Hoss, Thanks a lot. Will look into it. Regards, Eswar On Nov 26, 2007 11:55 PM, Chris Hostetter [EMAIL PROTECTED] wrote: : Does Solr come with Language analyzers for CJK? If not, can you please : direct me to some good CJK analyzers? Lucene has a CJKTokenizer and CJKAnalyzer in the contrib/analyzers jar. they can be used in Solr. both have been included in Solr for a while now, so you can specify CJKAnalyzer in your schema with Solr 1.2, but starting with Solr 1.3 a Factory for the Tokenizer will also be included so it can be used in a more complex analysis chain defined in the schema. -Hoss -- regards jl
Re: CJK Analyzers for Solr
i not use HYLANDA analyzer. i use je-analyzer and indexing at least 18m docs. i m sorry i only use chinese analyzer. On Nov 27, 2007 10:01 AM, Eswar K [EMAIL PROTECTED] wrote: What is the performance of these CJK analyzers (one in lucene and hylanda )? We would potentially be indexing millions of documents. James, We would have a look at hylanda too. What abt japanese and korean analyzers, any recommendations? - Eswar On Nov 27, 2007 7:21 AM, James liu [EMAIL PROTECTED] wrote: I don't think NGram is good method for Chinese. CJKAnalyzer of Lucene is 2-Gram. Eswar K: if it is chinese analyzer,,i recommend hylanda(www.hylanda.com),,,it is the best chinese analyzer and it not free. if u wanna free chinese analyzer, maybe u can try je-analyzer. it have some problem when using it. On Nov 27, 2007 5:56 AM, Otis Gospodnetic [EMAIL PROTECTED] wrote: Eswar, We've uses the NGram stuff that exists in Lucene's contrib/analyzers instead of CJK. Doesn't that allow you to do everything that the Chinese and CJK analyzers do? It's been a few months since I've looked at Chinese and CJK Analzyers, so I could be off. Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch - Original Message From: Eswar K [EMAIL PROTECTED] To: solr-user@lucene.apache.org Sent: Monday, November 26, 2007 8:30:52 AM Subject: CJK Analyzers for Solr Hi, Does Solr come with Language analyzers for CJK? If not, can you please direct me to some good CJK analyzers? Regards, Eswar -- regards jl -- regards jl
Re: Does SOLR supports multiple instances within the same webapplication?
if I understand correct,,u just do it like that:(i use php) $data1 = getDataFromInstance1($url); $data2 = getDataFromInstance2($url); it just have multi solr Instance. and getData from the distance. On Nov 12, 2007 11:15 PM, Dilip.TS [EMAIL PROTECTED] wrote: Hello, Does SOLR supports multiple instances within the same web application? If so how is this achieved? Thanks in advance. Regards, Dilip TS -- regards jl
Re: how to use PHP AND PHPS?
hmm i find error,,,that is my error not about php and phps .. i use old config to testso config have a problem.. that is Title i use double as its type...it should use text. On Nov 8, 2007 10:29 AM, James liu [EMAIL PROTECTED] wrote: php now is ok.. but phps failed mycode: ?php $url = 'http://localhost:8080/solr1/select/?q=2version=2.2rows=2fl=Titlestart=0rows=10indent=onwt=phps '; $a = file_get_contents($url); //eval('$solrResults = ' .$serializedSolrResults . ';'); echo 'bbefore unserialize/bbr/'; var_dump($a); echo 'Br/'; $a = unserialize($a); echo 'bafter unserialize.../bbr/'; var_dump($a); ? and result: *before unserialize* string(434) a:2:{s:14:responseHeader;a:3:{s:6:status;i:0;s:5:QTime;i:32;s:6:params;a:7:{s:2:fl;s:5:Title;s:6:indent;s:2:on;s:5:start;s:1:0;s:1:q;s:1:2;s:2:wt;s:4:phps;s:4:rows;a:2:{i:0;s:1:2;i:1;s:2:10;}s:7:version;s:3: 2.2;}}s:8:response;a:3:{s:8:numFound;i:28;s:5:start;i:0;s:4:docs;a:2:{i:0;a:1:{s:5:Title;d:诺基亚N-Gage基本数据;}i:1;a:1:{s:5:Title;d:索尼爱立信P908基本数据; *after unserialize...* bool(false) On Nov 7, 2007 9:30 PM, Dave Lewis [EMAIL PROTECTED] wrote: On Nov 7, 2007, at 2:04 AM, James liu wrote: i just decrease answer information...and u will see my result(full, not part) *before unserialize* string(433) a:2:{s:14:responseHeader;a:3:{s:6:status;i:0;s:5:QTime;i: 0;s:6:params;a:7:{s:2:fl;s:5:Title;s:6:indent;s:2:on;s: 5:start;s:1:0;s:1:q;s:1:2;s:2:wt;s:4:phps;s:4:rows;a: 2:{i:0;s:1:2;i:1;s:2:10;}s:7:version;s:3: 2.2;}}s:8:response;a:3:{s:8:numFound;i:28;s:5:start;i:0;s: 4:docs;a:2:{i:0;a:1:{s:5:Title;d:诺基亚N-Gage基本数据;}i:1;a:1: {s:5:Title;d:索尼爱立信P908基本数据; *after unserialize...* bool(false) and i write serialize test code.. ?php $ar = array ( array('id' = 123, 'Title'= 中文测试), array('id' = 123, 'Title'= 中国上海), ); echo serialize($ar); ? and result is : a:2:{i:0;a:2:{s:2:id;i:123;s:5:Title;s:12:中文测试;}i:1;a:2: {s:2:id;i:123;s:5:Title;s:12:中国上海;}} *php* result is: string(369) array( 'responseHeader'=array( 'status'=0, 'QTime'=0, 'params'=array( 'fl'='Title', 'indent'='on', 'start'='0', 'q'='2', 'wt'='php', 'rows'=array('2', '10'), 'version'='2.2')), 'response'=array('numFound'=28,'start'=0,'docs'=array( array( 'Title'=诺基亚N-Gage基本数据), array( 'Title'=索尼爱立信P908基本数 据)) )) it is string, so i can't read it correctly by php. This part (after string(369)) is exactly what it you should be seeing if you use the php handler, and it's what you get after you unserialize when using phps. You can access your search results as: $solrResults['response']['docs']; In your example above, that would be: array( array('Title'=诺基亚N-Gage基本数据), array( 'Title'=索尼爱立信 P908基本数据)) When using the php handler, you must do something like this: eval('$solrResults = ' .$serializedSolrResults . ';'); Then, as above, you can access $solrResults['response']['docs']. To sum up, if you use phps, you must unserialize the results. If you use php, you must eval the results (including some sugar to get a variable set to that value). dave -- regards jl -- regards jl
Re: how to use PHP AND PHPS?
php now is ok.. but phps failed mycode: ?php $url = ' http://localhost:8080/solr1/select/?q=2version=2.2rows=2fl=Titlestart=0rows=10indent=onwt=phps '; $a = file_get_contents($url); //eval('$solrResults = ' .$serializedSolrResults . ';'); echo 'bbefore unserialize/bbr/'; var_dump($a); echo 'Br/'; $a = unserialize($a); echo 'bafter unserialize.../bbr/'; var_dump($a); ? and result: *before unserialize* string(434) a:2:{s:14:responseHeader;a:3:{s:6:status;i:0;s:5:QTime;i:32;s:6:params;a:7:{s:2:fl;s:5:Title;s:6:indent;s:2:on;s:5:start;s:1:0;s:1:q;s:1:2;s:2:wt;s:4:phps;s:4:rows;a:2:{i:0;s:1:2;i:1;s:2:10;}s:7:version;s:3: 2.2;}}s:8:response;a:3:{s:8:numFound;i:28;s:5:start;i:0;s:4:docs;a:2:{i:0;a:1:{s:5:Title;d:诺基亚N-Gage基本数据;}i:1;a:1:{s:5:Title;d:索尼爱立信P908基本数据; *after unserialize...* bool(false) On Nov 7, 2007 9:30 PM, Dave Lewis [EMAIL PROTECTED] wrote: On Nov 7, 2007, at 2:04 AM, James liu wrote: i just decrease answer information...and u will see my result(full, not part) *before unserialize* string(433) a:2:{s:14:responseHeader;a:3:{s:6:status;i:0;s:5:QTime;i: 0;s:6:params;a:7:{s:2:fl;s:5:Title;s:6:indent;s:2:on;s: 5:start;s:1:0;s:1:q;s:1:2;s:2:wt;s:4:phps;s:4:rows;a: 2:{i:0;s:1:2;i:1;s:2:10;}s:7:version;s:3: 2.2;}}s:8:response;a:3:{s:8:numFound;i:28;s:5:start;i:0;s: 4:docs;a:2:{i:0;a:1:{s:5:Title;d:诺基亚N-Gage基本数据;}i:1;a:1: {s:5:Title;d:索尼爱立信P908基本数据; *after unserialize...* bool(false) and i write serialize test code.. ?php $ar = array ( array('id' = 123, 'Title'= 中文测试), array('id' = 123, 'Title'= 中国上海), ); echo serialize($ar); ? and result is : a:2:{i:0;a:2:{s:2:id;i:123;s:5:Title;s:12:中文测试;}i:1;a:2: {s:2:id;i:123;s:5:Title;s:12:中国上海;}} *php* result is: string(369) array( 'responseHeader'=array( 'status'=0, 'QTime'=0, 'params'=array( 'fl'='Title', 'indent'='on', 'start'='0', 'q'='2', 'wt'='php', 'rows'=array('2', '10'), 'version'='2.2')), 'response'=array('numFound'=28,'start'=0,'docs'=array( array( 'Title'=诺基亚N-Gage基本数据), array( 'Title'=索尼爱立信P908基本数 据)) )) it is string, so i can't read it correctly by php. This part (after string(369)) is exactly what it you should be seeing if you use the php handler, and it's what you get after you unserialize when using phps. You can access your search results as: $solrResults['response']['docs']; In your example above, that would be: array( array('Title'=诺基亚N-Gage基本数据), array( 'Title'=索尼爱立信 P908基本数据)) When using the php handler, you must do something like this: eval('$solrResults = ' .$serializedSolrResults . ';'); Then, as above, you can access $solrResults['response']['docs']. To sum up, if you use phps, you must unserialize the results. If you use php, you must eval the results (including some sugar to get a variable set to that value). dave -- regards jl
Re: how to use PHP AND PHPS?
first var_dump result(part not all): string(50506) a:2:{s:14:responseHeader;a:3:{s:6:status;i:0;s:5:QTime;i:2906;s:6:params;a:6:{s:6:indent;s:2:on;s:5:start;s:1:0;s:1:q;s:1:2;s:2:wt;s:4:phps;s:4:rows;s:2:10;s:7:version;s:3: 2.2;}} two var_dump result: bool(false) On Nov 6, 2007 10:36 PM, Dave Lewis [EMAIL PROTECTED] wrote: What are the results of the two var_dumps? dave On Nov 5, 2007, at 10:06 PM, James liu wrote: first: i m sure i enable php and phps in my solrconfig.xml two: i can't get answer. *phps: *?php $url = ' http://localhost:8080/solr1/select/? q=2version=2.2start=0rows=10indent=onwt=phps '; $a = file_get_contents($url); echo 'before unserializebr/'; var_dump($a); $a = unserialize($a); echo 'after unserialize...br/'; var_dump($a); ?* * On 11/6/07, Stu Hood [EMAIL PROTECTED] wrote: Did you enable the PHP serialized response writer in your solrconfig.xml? It is not enabled by default. Thanks, Stu -Original Message- From: James liu [EMAIL PROTECTED] Sent: Monday, November 5, 2007 9:03am To: solr-user@lucene.apache.org Subject: Re: how to use PHP AND PHPS? i know it...but u try it,,u will find simlar question. On 11/5/07, Robert Young [EMAIL PROTECTED] wrote: I would imagine you have to unserialize On 11/5/07, James liu [EMAIL PROTECTED] wrote: i find they all return string ?php $url = ' http://localhost:8080/solr/select/? q=solrversion=2.2start=0rows=10indent=onwt=php '; var_dump(file_get_contents($url); ? -- regards jl -- regards jl -- regards jl -- regards jl
Re: how to use PHP AND PHPS?
afternoon,,i will update svn...and try the newest... On Nov 7, 2007 11:23 AM, Dave Lewis [EMAIL PROTECTED] wrote: On Nov 6, 2007, at 8:10 PM, James liu wrote: first var_dump result(part not all): string(50506) a:2:{s:14:responseHeader;a:3:{s:6:status;i:0;s:5:QTime;i: 2906;s:6:params;a:6:{s:6:indent;s:2:on;s:5:start;s:1:0;s: 1:q;s:1:2;s:2:wt;s:4:phps;s:4:rows;s:2:10;s: 7:version;s:3: 2.2;}} This is exactly correct. two var_dump result: bool(false) So, unserializing is failing. Are you running from the trunk or from a nightly? There was a bug a couple of weeks ago that sent back faulty serialized data. It's fixed now. It's possible this is your issue. dave On Nov 6, 2007 10:36 PM, Dave Lewis [EMAIL PROTECTED] wrote: What are the results of the two var_dumps? dave On Nov 5, 2007, at 10:06 PM, James liu wrote: first: i m sure i enable php and phps in my solrconfig.xml two: i can't get answer. *phps: *?php $url = ' http://localhost:8080/solr1/select/? q=2version=2.2start=0rows=10indent=onwt=phps '; $a = file_get_contents($url); echo 'before unserializebr/'; var_dump($a); $a = unserialize($a); echo 'after unserialize...br/'; var_dump($a); ?* * On 11/6/07, Stu Hood [EMAIL PROTECTED] wrote: Did you enable the PHP serialized response writer in your solrconfig.xml? It is not enabled by default. Thanks, Stu -Original Message- From: James liu [EMAIL PROTECTED] Sent: Monday, November 5, 2007 9:03am To: solr-user@lucene.apache.org Subject: Re: how to use PHP AND PHPS? i know it...but u try it,,u will find simlar question. On 11/5/07, Robert Young [EMAIL PROTECTED] wrote: I would imagine you have to unserialize On 11/5/07, James liu [EMAIL PROTECTED] wrote: i find they all return string ?php $url = ' http://localhost:8080/solr/select/? q=solrversion=2.2start=0rows=10indent=onwt=php '; var_dump(file_get_contents($url); ? -- regards jl -- regards jl -- regards jl -- regards jl -- regards jl
Re: how to use PHP AND PHPS?
same answer. On Nov 7, 2007 11:41 AM, James liu [EMAIL PROTECTED] wrote: afternoon,,i will update svn...and try the newest... On Nov 7, 2007 11:23 AM, Dave Lewis [EMAIL PROTECTED] wrote: On Nov 6, 2007, at 8:10 PM, James liu wrote: first var_dump result(part not all): string(50506) a:2:{s:14:responseHeader;a:3:{s:6:status;i:0;s:5:QTime;i: 2906;s:6:params;a:6:{s:6:indent;s:2:on;s:5:start;s:1:0;s: 1:q;s:1:2;s:2:wt;s:4:phps;s:4:rows;s:2:10;s: 7:version;s:3: 2.2;}} This is exactly correct. two var_dump result: bool(false) So, unserializing is failing. Are you running from the trunk or from a nightly? There was a bug a couple of weeks ago that sent back faulty serialized data. It's fixed now. It's possible this is your issue. dave On Nov 6, 2007 10:36 PM, Dave Lewis [EMAIL PROTECTED] wrote: What are the results of the two var_dumps? dave On Nov 5, 2007, at 10:06 PM, James liu wrote: first: i m sure i enable php and phps in my solrconfig.xml two: i can't get answer. *phps: *?php $url = ' http://localhost:8080/solr1/select/? q=2version= 2.2start=0rows=10indent=onwt=phps '; $a = file_get_contents($url); echo 'before unserializebr/'; var_dump($a); $a = unserialize($a); echo 'after unserialize...br/'; var_dump($a); ?* * On 11/6/07, Stu Hood [EMAIL PROTECTED] wrote: Did you enable the PHP serialized response writer in your solrconfig.xml? It is not enabled by default. Thanks, Stu -Original Message- From: James liu [EMAIL PROTECTED] Sent: Monday, November 5, 2007 9:03am To: solr-user@lucene.apache.org Subject: Re: how to use PHP AND PHPS? i know it...but u try it,,u will find simlar question. On 11/5/07, Robert Young [EMAIL PROTECTED] wrote: I would imagine you have to unserialize On 11/5/07, James liu [EMAIL PROTECTED] wrote: i find they all return string ?php $url = ' http://localhost:8080/solr/select/? q=solrversion=2.2start=0rows=10indent=onwt=php '; var_dump(file_get_contents($url); ? -- regards jl -- regards jl -- regards jl -- regards jl -- regards jl -- regards jl
Re: how to use PHP AND PHPS?
i just decrease answer information...and u will see my result(full, not part) *before unserialize* string(433) a:2:{s:14:responseHeader;a:3:{s:6:status;i:0;s:5:QTime;i:0;s:6:params;a:7:{s:2:fl;s:5:Title;s:6:indent;s:2:on;s:5:start;s:1:0;s:1:q;s:1:2;s:2:wt;s:4:phps;s:4:rows;a:2:{i:0;s:1:2;i:1;s:2:10;}s:7:version;s:3: 2.2;}}s:8:response;a:3:{s:8:numFound;i:28;s:5:start;i:0;s:4:docs;a:2:{i:0;a:1:{s:5:Title;d:诺基亚N-Gage基本数据;}i:1;a:1:{s:5:Title;d:索尼爱立信P908基本数据; *after unserialize...* bool(false) and i write serialize test code.. ?php $ar = array ( array('id' = 123, 'Title'= 中文测试), array('id' = 123, 'Title'= 中国上海), ); echo serialize($ar); ? and result is : a:2:{i:0;a:2:{s:2:id;i:123;s:5:Title;s:12:中文测试;}i:1;a:2:{s:2:id;i:123;s:5:Title;s:12:中国上海;}} *php* result is: string(369) array( 'responseHeader'=array( 'status'=0, 'QTime'=0, 'params'=array( 'fl'='Title', 'indent'='on', 'start'='0', 'q'='2', 'wt'='php', 'rows'=array('2', '10'), 'version'='2.2')), 'response'=array('numFound'=28,'start'=0,'docs'=array( array( 'Title'=诺基亚N-Gage基本数据), array( 'Title'=索尼爱立信P908基本数据)) )) it is string, so i can't read it correctly by php. -- regards jl
how to use PHP AND PHPS?
i find they all return string ?php $url = ' http://localhost:8080/solr/select/?q=solrversion=2.2start=0rows=10indent=onwt=php '; var_dump(file_get_contents($url); ? -- regards jl
Re: how to use PHP AND PHPS?
first: i m sure i enable php and phps in my solrconfig.xml two: i can't get answer. *phps: *?php $url = ' http://localhost:8080/solr1/select/?q=2version=2.2start=0rows=10indent=onwt=phps '; $a = file_get_contents($url); echo 'before unserializebr/'; var_dump($a); $a = unserialize($a); echo 'after unserialize...br/'; var_dump($a); ?* * On 11/6/07, Stu Hood [EMAIL PROTECTED] wrote: Did you enable the PHP serialized response writer in your solrconfig.xml? It is not enabled by default. Thanks, Stu -Original Message- From: James liu [EMAIL PROTECTED] Sent: Monday, November 5, 2007 9:03am To: solr-user@lucene.apache.org Subject: Re: how to use PHP AND PHPS? i know it...but u try it,,u will find simlar question. On 11/5/07, Robert Young [EMAIL PROTECTED] wrote: I would imagine you have to unserialize On 11/5/07, James liu [EMAIL PROTECTED] wrote: i find they all return string ?php $url = ' http://localhost:8080/solr/select/?q=solrversion=2.2start=0rows=10indent=onwt=php '; var_dump(file_get_contents($url); ? -- regards jl -- regards jl -- regards jl
Re: get all query not working
if u rebuild solr , safe method is rm -r *tomcat*/webapps/*. 2007/11/1, Chris Hostetter [EMAIL PROTECTED]: : Is there an easy to find out which version of solr is running. I installed : solr 1.2 and set up an instance using Tomcat. It was successful before. FYI: starting a while back, the Info page (registry.jsp) of the admin interface gives you specifics on the Solr and Lucene versions in use. -Hoss -- regards jl
Re: SOLR 1.3 Release?
where i can read 1.3 new features? 2007/10/26, Venkatraman S [EMAIL PROTECTED]: On 10/26/07, Mike Klaas [EMAIL PROTECTED] wrote: If we did a 1.2.x, it shoud (imo) contain no new features, only important bugfixes. I have been having a look at the trunk for quite sometime now, and must say that its changing pretty fast. Having an interim release now will require more work with the comitters having to put an extra effort to spruce up things. I would personally suggest 1-jan-2008 as the next Solr release which would help in sprucing up the code and also releasing the bunch of new features that have been lying in the trunk to be used in *the-next-stable-release*. -Venkat -- Blog @ http://blizzardzblogs.blogspot.com -- regards jl
random lost segments* files,,
i have 40 instances,,,one instance lost segments* file(happen after commit and optimize) anyone have similar problem? can i fix this problem? can i recovery this instance data? -- regards jl
Re: random lost segments* files,,
i find it happen when it do commit. i use solr 1.2 release. i use crontab to do index work. 2007/10/15, James liu [EMAIL PROTECTED]: i have 40 instances,,,one instance lost segments* file(happen after commit and optimize) anyone have similar problem? can i fix this problem? can i recovery this instance data? -- regards jl -- regards jl
Re: Availability Issues
* field name=text type=text indexed=true stored=true multiValued=true/* *i think text not need stored='true' unless u will show it.(it will help u decrease index size and not affect search )* *index and search use same box? if it is true, u should moniter search response time when indexing.(include CPU, RAM change)* *i have similar problem and i increase JVM size to fix it.(u can try it and show me your response)* 2007/10/9, David Whalen [EMAIL PROTECTED]: Thanks for letting me know that. Okay, here they are: BEGIN SCHEMA.XML=== ?xml version=1.0 ? !-- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the License); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an AS IS BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. -- !-- This is the Solr schema file. This file should be named schema.xml and should be in the conf directory under the solr home (i.e. ./solr/conf/schema.xml by default) or located where the classloader for the Solr webapp can find it. For more information, on how to customize this file, please see http://wiki.apache.org/solr/SchemaXml -- schema name=enr-solr version=1.1 !-- attribute name is the name of this schema and is only used for display purposes. Applications should change this to reflect the nature of the search collection. version=1.1 is Solr's version number for the schema syntax and semantics. It should not normally be changed by applications. 1.0: multiValued attribute did not exist, all fields are multiValued by nature 1.1: multiValued attribute introduced, false by default -- types !-- field type definitions. The name attribute is just a label to be used by field definitions. The class attribute and any other attributes determine the real behavior of the fieldtype. Class names starting with solr refer to java classes in the org.apache.solr.analysis package. -- !-- The StrField type is not analyzed, but indexed/stored verbatim. - StrField and TextField support an optional compressThreshold which limits compression (if enabled in the derived fields) to values which exceed a certain size (in characters). -- fieldtype name=string class=solr.StrField sortMissingLast=true omitNorms=true/ !-- boolean type: true or false -- fieldtype name=boolean class=solr.BoolField sortMissingLast=true omitNorms=true/ !-- The optional sortMissingLast and sortMissingFirst attributes are currently supported on types that are sorted internally as strings. - If sortMissingLast=true, then a sort on this field will cause documents without the field to come after documents with the field, regardless of the requested sort order (asc or desc). - If sortMissingFirst=true, then a sort on this field will cause documents without the field to come before documents with the field, regardless of the requested sort order. - If sortMissingLast=false and sortMissingFirst=false (the default), then default lucene sorting will be used which places docs without the field first in an ascending sort and last in a descending sort. -- !-- numeric field types that store and index the text value verbatim (and hence don't support range queries, since the lexicographic ordering isn't equal to the numeric ordering) -- fieldtype name=integer class=solr.IntField omitNorms=true/ fieldtype name=long class=solr.LongField omitNorms=true/ fieldtype name=float class=solr.FloatField omitNorms=true/ fieldtype name=double class=solr.DoubleField omitNorms=true/ !-- Numeric field types that manipulate the value into a string value that isn't human-readable in its internal form, but with a lexicographic ordering the same as the numeric ordering, so that range queries work correctly. -- fieldtype name=sint class=solr.SortableIntField sortMissingLast=true omitNorms=true/ fieldtype name=slong class=solr.SortableLongField sortMissingLast=true omitNorms=true/ fieldtype name=sfloat class=solr.SortableFloatField sortMissingLast=true omitNorms=true/ fieldtype name=sdouble class=solr.SortableDoubleField sortMissingLast=true omitNorms=true/ !-- The format for
anyone can send me jetty-plus
i can't download it from http://jetty.mortbay.org/jetty5/plus/index.html -- regards jl
Re: real time search
if new data come in and drive index it, load new search it. if more docs, optimize time will cost much, so can't do search like real time. so i think new solr instance only for newest information. the docs will be ~10K. if it arrive 10k, it should be closed and rebuild new instance. (if we have more memory and cpu,,,just open more instance ,not one by one.) final, we merge result and show it to user. that all i think, not test it. 2007/9/24, Grant Ingersoll [EMAIL PROTECTED]: Hi James, Can you provide more information about what you are trying to do? By real time search, do you mean you want indexed documents to be available immediately? Or is a minute or two acceptable? Do all users need to see them immediately, or just the current user? We can better help you if you give us more details on what you are trying to accomplish. Thanks, Grant On Sep 23, 2007, at 10:38 AM, James liu wrote: i wanna do it. Maybe someone did it, if so, give me some tips. thks -- regards jl -- Grant Ingersoll http://lucene.grantingersoll.com Lucene Helpful Hints: http://wiki.apache.org/lucene-java/BasicsOfPerformance http://wiki.apache.org/lucene-java/LuceneFAQ -- regards jl
can solr do it?
if use multi solr with one index, it will cache individually. so i think can it share their cache.(they have same config) -- regards jl
Re: real time search
it seems use somthing like ajax... if so, it not what i wanna 2007/9/25, Matthew Runo [EMAIL PROTECTED]: I assume you mean something like this: http://addictedtonew.com/archives/145/wordpress-live-search-plugin/ Take a look at how the search box works - is that what you mean? ++ | Matthew Runo | Zappos Development | [EMAIL PROTECTED] | 702-943-7833 ++ On Sep 24, 2007, at 8:13 AM, Grant Ingersoll wrote: Hi James, Can you provide more information about what you are trying to do? By real time search, do you mean you want indexed documents to be available immediately? Or is a minute or two acceptable? Do all users need to see them immediately, or just the current user? We can better help you if you give us more details on what you are trying to accomplish. Thanks, Grant On Sep 23, 2007, at 10:38 AM, James liu wrote: i wanna do it. Maybe someone did it, if so, give me some tips. thks -- regards jl -- Grant Ingersoll http://lucene.grantingersoll.com Lucene Helpful Hints: http://wiki.apache.org/lucene-java/BasicsOfPerformance http://wiki.apache.org/lucene-java/LuceneFAQ -- regards jl
real time search
i wanna do it. Maybe someone did it, if so, give me some tips. thks -- regards jl
How to patch
i wanna try patch: https://issues.apache.org/jira/browse/SOLR-139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel and i download solr1.2 release patch SOLR-269*.pach(when in '/tmp/apache-solr-1.2.0/src/test/org/apache/solr/update' ) it show me |Index: src/test/org/apache/solr/common/SolrDocumentTest.java |=== |--- src/test/org/apache/solr/common/SolrDocumentTest.java (revision 555897) |+++ src/test/org/apache/solr/common/SolrDocumentTest.java (working copy) -- File to patch: i input the absolute path which file should be ... but it not ok... does my solr's version not right? -- regards jl
Re: Multiple indexes
OK...I see...thk u ,mike. 2007/8/31, Mike Klaas [EMAIL PROTECTED]: On 29-Aug-07, at 10:21 PM, James liu wrote: Does it affect with doc size? for example 2 billion docs, 10k doc2 billion docs, but doc size is 10m. There might be other places that have 2G limit (see lucene index format docs), but many things are vints and can grow larger. Of course you will hit physical limits of your machine long before you can achieve your hypothetical situation: that's 20,000 Tb, which is many, many times the size of a complete internet crawl. -Mike 2007/8/30, Mike Klaas [EMAIL PROTECTED]: 2 billion docs (signed int). On 29-Aug-07, at 6:24 PM, James liu wrote: what is the limits for Lucene and Solr. 100m, 1000m, 5000m or other number docs? 2007/8/24, Walter Underwood [EMAIL PROTECTED]: It should work fine to index them and search them. 13 million docs is not even close to the limits for Lucene and Solr. Have you had problems? wunder On 8/23/07 7:30 AM, Jae Joo [EMAIL PROTECTED] wrote: Is there any solution to handle 13 millions document shown as below? Each document is not big, but the number of ones is 13 million. Any way to utilize the multiple indexes? Thanks, Jae Joo docfield name=trade2/field field name=company_nameUnlimi-Tech Software Inc/field field name=phys_stabrvON/field field name=trade4/field field name=status_id_descrSingle Location/field field name=trade3/field field name=phys_countryCanada/field field name=phys_zipK1C 4R1/field field name=phys_cityOttawa/field field name=phys_stateOntario/field field name=sic2G2_Computer Software/field field name=phys_address1447a Youville Dr/field field name=sic1G_Technology amp; Communications/field field name=duns_number203439018/field field name=trade1/field field name=phys_countyCarleton/field field name=trade5/field field name=status_id_rank30/field field name=sic4G2173_Computer Programming Services/field field name=sic8G217308D_Computer software development/field /doc -- regards jl -- regards jl -- regards jl
Re: Multiple indexes
Does it affect with doc size? for example 2 billion docs, 10k doc2 billion docs, but doc size is 10m. 2007/8/30, Mike Klaas [EMAIL PROTECTED]: 2 billion docs (signed int). On 29-Aug-07, at 6:24 PM, James liu wrote: what is the limits for Lucene and Solr. 100m, 1000m, 5000m or other number docs? 2007/8/24, Walter Underwood [EMAIL PROTECTED]: It should work fine to index them and search them. 13 million docs is not even close to the limits for Lucene and Solr. Have you had problems? wunder On 8/23/07 7:30 AM, Jae Joo [EMAIL PROTECTED] wrote: Is there any solution to handle 13 millions document shown as below? Each document is not big, but the number of ones is 13 million. Any way to utilize the multiple indexes? Thanks, Jae Joo docfield name=trade2/field field name=company_nameUnlimi-Tech Software Inc/field field name=phys_stabrvON/field field name=trade4/field field name=status_id_descrSingle Location/field field name=trade3/field field name=phys_countryCanada/field field name=phys_zipK1C 4R1/field field name=phys_cityOttawa/field field name=phys_stateOntario/field field name=sic2G2_Computer Software/field field name=phys_address1447a Youville Dr/field field name=sic1G_Technology amp; Communications/field field name=duns_number203439018/field field name=trade1/field field name=phys_countyCarleton/field field name=trade5/field field name=status_id_rank30/field field name=sic4G2173_Computer Programming Services/field field name=sic8G217308D_Computer software development/field /doc -- regards jl -- regards jl
what's the difference between lucene2.1 and solr1.2
if i wanna calc it by my method, something i should notice ? anyone did it? -- regards jl
why store field will be analyzed?
fieldset topic indexed='false' and stored='true' i don't know why it will be analyzed? now i wanna it only store not analyzed,,,how can i do? -- regards jl
Re: solr index problem
I correct it,,,i index 17M docs. not 1.7M,,,so OutOfMemory happen when it finish index ~11.3m docs It is new index. i think it maybe the reason: On 7/18/07, Otis Gospodnetic [EMAIL PROTECTED] wrote: Why? Too small of a Java heap. :) Increase the size of the Java heap and lower the maxBufferedDocs number in solrconfig.xml and then try again. If it only happens after a lot of docs, it's probably not maxBufferedDocs, but when a big luicene merge is triggered. 2007/7/19, Mike Klaas [EMAIL PROTECTED]: On 18-Jul-07, at 2:58 PM, Yonik Seeley wrote: On 7/18/07, Mike Klaas [EMAIL PROTECTED] wrote: Could happen when doDeleting the pending docs too. James: try sending commit every 500k docs or so. Hmmm, right... some of the memory usage will be related to the treemap keeping track of deleted items, and some of it will be related to the reader itself (the term index, and the norms). Perhaps we should have some sort of limit on the number of deletions we cache so people don't have to worry about that part. I logged this in http://issues.apache.org/jira/browse/SOLR-310 Perhaps a solrconfig.xml setting, maxPendingDocs? Default to 100k? I have experienced this from time to time... it is one of the reasons I fastidiously wipe the index and restart solr before reindexing. aha,,me too。 -Mike -- regards jl
solr index problem
when i index 1.7m docs and 4k-5k per doc. OutOfMemory happen when it finish index ~1.13m docs I just restart tomcat , delete all lock and restart do index. No error or warning infor until it finish. anyone know why? or have the same error? -- regards jl
Re: Solr Server Configuration
u can find configuration datadir in solrconfig.xml(solr 1.2) 2007/7/10, nithyavembu [EMAIL PROTECTED]: Hi, I tried as you said and got the result without any error. So we can make the solr home anywhere. But we have to give the path correctly in solr.xml . Am i correct? Now i am one step further.. :) Best, V.Nithya. Kijiji Xu, Ping wrote: I'm sorry I don't quite understand why you create a folder named solr in tomcat/bin in your step 3 The first of all ,you should understand this solr.xml file which I put in tomcat/conf/Catalina/localhost/,its contents follows like this: Context docBase=/home/tomcat/webapps/solr.war debug=0 crossContext=true Environment name=solr/home type=java.lang.String value=/home/solr override=true / /Context When tomcat was start,tomcat will read this file,and then,tomcat know that he should unzip the /home/tomcat/webapps/solr.war,deploy it ,you'll see a folder named solr in the tomcat webapps folder when tomcat be started at last. And Environment name=solr/home type=java.lang.String value=/home/solr override=true / means solr's home path is /home/solr, in the folder '/home/solr',it contains folder 'conf' with solr's configuration file:schema.xml and solrconfig.xml ,so at last file structure will be like this: /home/tomcat/conf/Catalina/localhost/solr.xml /home/tomcat/webapps/solr.war /home/solr/conf/schema.xml /home/solr/conf/solrconfig.xml I attacked a home.rar file in this mail,it is solr's files in the tomcat directory structure,FYI My msn: [EMAIL PROTECTED] if there is problem,pls contact with me. -Original Message- From: nithyavembu [mailto:[EMAIL PROTECTED] Sent: 2007年7月10日 17:14 To: solr-user@lucene.apache.org Subject: RE: Solr Server Configuration Hi Kijiji Xu, Thanks a lot for your clear guidance. I tried as you said. But i was unclear with the environment varible values in solr.xml So atlast i didnt touch the configuration parts. These are the steps i followed for the solr server working.. 1. Used tomcat 5 with windows. 2. Deployed the war apache-solr-1.2.0 in to the webapp directory and rename it as solr.war 3. Then i created one folder named solr in tomcat - bin. 4. Copy the conf folder in example folder and placed that conf in the solr folder in tomcat bin. 5. Then start the server and and tried the url http://localhost:8080/solr/ in browser. Is there anything wrong i did? If so please tell me.. with Regards, V.Nithya. Kijiji Xu, Ping wrote: i) In lucene we add document by using java. But here they given in xml format and posting them using POST.jar. How can i add document in java? In java ,there have a class named org.apache.solr.util.SimplePostTool ,you can look at the source code of it,and you'll know how it worked for posting xml data to solr ii) How can i send the http request n get response in a standalone java program? In my opinion, you can use the java HttpURLConnection class to request and get response from the solr server iii) Whether i have to add any xml or i have to modify the xml (server.xml,web.xml) files for solr server configuration? NO, but there have two type of configuration for solr . First of all, the solr configuration file locate in the solr/conf,there are solrconfig.xml and schema.xml ,you would change them for your application . Second, you have to deploy solr application itself to Tomcat, and you would put a configuration file like this in the tomcat ,the path is /home/tomcat/conf/Catalina/localhost/: Context docBase=/home/tomcat/webapps/solr.war debug=0 crossContext=true Environment name=solr/home type=java.lang.String value=/home/tomcat/solr override=true / /Context That's all,if you have any question,feel free to ask me:) -Original Message- From: nithyavembu [mailto:[EMAIL PROTECTED] Sent: 2007年7月6日 13:27 To: solr-user@lucene.apache.org Subject: Solr Server Configuration Hi All, I am new to Solr. But i am familiar with lucene. I am stuggling in solr server configuation.I am using Tomcat 5. I have worked with the example standalone given with solr while i download. But my doubt is i) In lucene we add document by using java. But here they given in xml format and posting them using POST.jar. How can i add document in java? ii) How can i send the http request n get response in a standalone java program? iii) Whether i have to add any xml or i have to modify the xml (server.xml,web.xml) files for solr server configuration? Best, Nithya.V. -- View this message in context: http://www.nabble.com/Solr-Server-Configuration-tf4033748.html#a11458997 Sent from the Solr - User mailing list archive at Nabble.com. -- View this message in context: http://www.nabble.com/Solr-Server-Configuration-tf4033748.html#a11517331 Sent from the Solr - User mailing list archive at Nabble.com. -- View this message in context:
Re: who use time?
I use freebsd. 2007/6/16, Yonik Seeley [EMAIL PROTECTED]: On 6/14/07, James liu [EMAIL PROTECTED] wrote: I just timing my script to get data from 2 solr boxes, not complete script. It just query two box and return id,score .rows=10. response type use json. and i see their qtime all zero. How many documents are you retrieving? one: numDocs : 1 maxDoc : 23000 the other: numDocs : 9000 maxDoc : 9000 3-4k per doc If you re-execute the same exact query again, is it still slower? It will be quick. time will be 0.0043279ms It's probably the time it takes to retrieve the stored fields from the index... perhaps the OS had to go to disk. -Yonik -- regards jl
Re: i wanna change response type to PHP serialize
Can u write a test and i give u chinese word. u can try it.. or u get chinese word from china web site..http://www.sina.com.cn,,, i don't know to send sample result to u. 2007/6/28, Nick Jenkin [EMAIL PROTECTED]: Are you able to attach a file of serialized (direct php serialized output from solr) result so I can test? Thanks -Nick On 6/28/07, James liu [EMAIL PROTECTED] wrote: code not change,,,and i not use utf8_decodeshould do it? 2007/6/28, Nick Jenkin [EMAIL PROTECTED]: Hi James It is totally not optimized, when you say change your content into ???, I assume this is because of UTF8 issues, are you using utf8_decode etc? Thanks -Nick On 6/28/07, James liu [EMAIL PROTECTED] wrote: It is slower than json and xml,,,and it will change my content into ??? when i use json , content is ok. afternoon, iwill read ur code. 2007/6/27, James liu [EMAIL PROTECTED]: ok,,thks nick,,,i just forget replace jar file.. wait a minute i will test speed... 2007/6/27, Nick Jenkin [EMAIL PROTECTED]: http://nickjenkin.com/misc/apache-solr-1.2.0-php-serialize.tar.gz Try that -Nick On 6/27/07, James liu [EMAIL PROTECTED] wrote: i use tomcat ,, send ur solr version to me...i try it again.. 2007/6/27, Nick Jenkin [EMAIL PROTECTED]: If you are using the example provided in 1.2 (using jetty) you need to use ant example rather than ant dist -Nick On 6/27/07, James liu [EMAIL PROTECTED] wrote: Yes, i use 1.2my compile method: download solr 1.2 and modify file by your patch. use `ant dist` to compile it. no error show. i can see its admin gui, but i wanna try search, it will show me the error information, Is my compile method right? if not, show me how to compile it. it seems very strange ,only me fail? anyone have same question? if free, maybe u zip your solr to me by mail...and i try it again. 2007/6/26, Nick Jenkin [EMAIL PROTECTED] : Interesting, what version of solr are you using, I tested on 1.2 . -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: i just cp it to src\java\org\apache\solr\request and ant dist...i think maybe my method is wrong. same error infromation... java.lang.NoClassDefFoundError: org/apache/solr/search/ScorePriorityQueue at org.apache.solr.search.SolrIndexSearcher.getDocListNC( SolrIndexSearcher.java:886) at org.apache.solr.search.SolrIndexSearcher.getDocListC ( SolrIndexSearcher.java:805) at org.apache.solr.search.SolrIndexSearcher.getDocList ( SolrIndexSearcher.java:698) at org.apache.solr.request.StandardRequestHandler.handleRequestBody ( StandardRequestHandler.java:122) at org.apache.solr.handler.RequestHandlerBase.handleRequest ( RequestHandlerBase.java:77) at org.apache.solr.core.SolrCore.execute( SolrCore.java :658) at org.apache.solr.servlet.SolrServlet.doGet ( SolrServlet.java :66) at javax.servlet.http.HttpServlet.service( HttpServlet.java :690) at javax.servlet.http.HttpServlet.service ( HttpServlet.java :803) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter ( ApplicationFilterChain.java:290) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.solr.servlet.SolrDispatchFilter.doFilter( SolrDispatchFilter.java:185) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter ( ApplicationFilterChain.java :235) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke ( StandardWrapperValve.java:228) at org.apache.catalina.core.StandardContextValve.invoke( StandardContextValve.java:175) at org.apache.catalina.core.StandardHostValve.invoke( StandardHostValve.java:128) at org.apache.catalina.valves.ErrorReportValve.invoke( ErrorReportValve.java :104) at org.apache.catalina.core.StandardEngineValve.invoke ( StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service( CoyoteAdapter.java
Re: i wanna change response type to PHP serialize
It is slower than json and xml,,,and it will change my content into ??? when i use json , content is ok. afternoon, iwill read ur code. 2007/6/27, James liu [EMAIL PROTECTED]: ok,,thks nick,,,i just forget replace jar file.. wait a minute i will test speed... 2007/6/27, Nick Jenkin [EMAIL PROTECTED]: http://nickjenkin.com/misc/apache-solr-1.2.0-php-serialize.tar.gz Try that -Nick On 6/27/07, James liu [EMAIL PROTECTED] wrote: i use tomcat ,, send ur solr version to me...i try it again.. 2007/6/27, Nick Jenkin [EMAIL PROTECTED]: If you are using the example provided in 1.2 (using jetty) you need to use ant example rather than ant dist -Nick On 6/27/07, James liu [EMAIL PROTECTED] wrote: Yes, i use 1.2my compile method: download solr 1.2 and modify file by your patch. use `ant dist` to compile it. no error show. i can see its admin gui, but i wanna try search, it will show me the error information, Is my compile method right? if not, show me how to compile it. it seems very strange ,only me fail? anyone have same question? if free, maybe u zip your solr to me by mail...and i try it again. 2007/6/26, Nick Jenkin [EMAIL PROTECTED] : Interesting, what version of solr are you using, I tested on 1.2 . -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: i just cp it to src\java\org\apache\solr\request and ant dist...i think maybe my method is wrong. same error infromation... java.lang.NoClassDefFoundError: org/apache/solr/search/ScorePriorityQueue at org.apache.solr.search.SolrIndexSearcher.getDocListNC( SolrIndexSearcher.java:886) at org.apache.solr.search.SolrIndexSearcher.getDocListC ( SolrIndexSearcher.java:805) at org.apache.solr.search.SolrIndexSearcher.getDocList ( SolrIndexSearcher.java:698) at org.apache.solr.request.StandardRequestHandler.handleRequestBody ( StandardRequestHandler.java:122) at org.apache.solr.handler.RequestHandlerBase.handleRequest ( RequestHandlerBase.java:77) at org.apache.solr.core.SolrCore.execute(SolrCore.java :658) at org.apache.solr.servlet.SolrServlet.doGet ( SolrServlet.java :66) at javax.servlet.http.HttpServlet.service( HttpServlet.java :690) at javax.servlet.http.HttpServlet.service ( HttpServlet.java :803) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter ( ApplicationFilterChain.java:290) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.solr.servlet.SolrDispatchFilter.doFilter( SolrDispatchFilter.java:185) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter ( ApplicationFilterChain.java :235) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke ( StandardWrapperValve.java:228) at org.apache.catalina.core.StandardContextValve.invoke( StandardContextValve.java:175) at org.apache.catalina.core.StandardHostValve.invoke( StandardHostValve.java:128) at org.apache.catalina.valves.ErrorReportValve.invoke( ErrorReportValve.java :104) at org.apache.catalina.core.StandardEngineValve.invoke ( StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service( CoyoteAdapter.java:216) at org.apache.coyote.http11.Http11Processor.process( Http11Processor.java:844) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process( Http11Protocol.java:634) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run ( JIoEndpoint.java:445) at java.lang.Thread.run(Unknown Source) 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I have some good news :o) https://issues.apache.org/jira/browse/SOLR-275 Please let me know if you find any bugs Thanks -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: I think it simple to u. so i wait for ur good news. 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I am also quite interested in getting a serialized PHP array response writer, after some investigation it doesn't seem as difficult as I first thought, I will have a try at implementing this when I get some time, the format of the array would probably end up being the same as if you were to use json_decode. -Nick
Re: i wanna change response type to PHP serialize
code not change,,,and i not use utf8_decodeshould do it? 2007/6/28, Nick Jenkin [EMAIL PROTECTED]: Hi James It is totally not optimized, when you say change your content into ???, I assume this is because of UTF8 issues, are you using utf8_decode etc? Thanks -Nick On 6/28/07, James liu [EMAIL PROTECTED] wrote: It is slower than json and xml,,,and it will change my content into ??? when i use json , content is ok. afternoon, iwill read ur code. 2007/6/27, James liu [EMAIL PROTECTED]: ok,,thks nick,,,i just forget replace jar file.. wait a minute i will test speed... 2007/6/27, Nick Jenkin [EMAIL PROTECTED]: http://nickjenkin.com/misc/apache-solr-1.2.0-php-serialize.tar.gz Try that -Nick On 6/27/07, James liu [EMAIL PROTECTED] wrote: i use tomcat ,, send ur solr version to me...i try it again.. 2007/6/27, Nick Jenkin [EMAIL PROTECTED]: If you are using the example provided in 1.2 (using jetty) you need to use ant example rather than ant dist -Nick On 6/27/07, James liu [EMAIL PROTECTED] wrote: Yes, i use 1.2my compile method: download solr 1.2 and modify file by your patch. use `ant dist` to compile it. no error show. i can see its admin gui, but i wanna try search, it will show me the error information, Is my compile method right? if not, show me how to compile it. it seems very strange ,only me fail? anyone have same question? if free, maybe u zip your solr to me by mail...and i try it again. 2007/6/26, Nick Jenkin [EMAIL PROTECTED] : Interesting, what version of solr are you using, I tested on 1.2 . -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: i just cp it to src\java\org\apache\solr\request and ant dist...i think maybe my method is wrong. same error infromation... java.lang.NoClassDefFoundError: org/apache/solr/search/ScorePriorityQueue at org.apache.solr.search.SolrIndexSearcher.getDocListNC( SolrIndexSearcher.java:886) at org.apache.solr.search.SolrIndexSearcher.getDocListC ( SolrIndexSearcher.java:805) at org.apache.solr.search.SolrIndexSearcher.getDocList ( SolrIndexSearcher.java:698) at org.apache.solr.request.StandardRequestHandler.handleRequestBody ( StandardRequestHandler.java:122) at org.apache.solr.handler.RequestHandlerBase.handleRequest ( RequestHandlerBase.java:77) at org.apache.solr.core.SolrCore.execute( SolrCore.java :658) at org.apache.solr.servlet.SolrServlet.doGet ( SolrServlet.java :66) at javax.servlet.http.HttpServlet.service( HttpServlet.java :690) at javax.servlet.http.HttpServlet.service ( HttpServlet.java :803) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter ( ApplicationFilterChain.java:290) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.solr.servlet.SolrDispatchFilter.doFilter( SolrDispatchFilter.java:185) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter ( ApplicationFilterChain.java :235) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke ( StandardWrapperValve.java:228) at org.apache.catalina.core.StandardContextValve.invoke( StandardContextValve.java:175) at org.apache.catalina.core.StandardHostValve.invoke( StandardHostValve.java:128) at org.apache.catalina.valves.ErrorReportValve.invoke( ErrorReportValve.java :104) at org.apache.catalina.core.StandardEngineValve.invoke ( StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service( CoyoteAdapter.java:216) at org.apache.coyote.http11.Http11Processor.process( Http11Processor.java:844) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process( Http11Protocol.java:634) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run ( JIoEndpoint.java:445) at java.lang.Thread.run(Unknown Source) 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I have some good news :o) https://issues.apache.org/jira/browse/SOLR-275 Please let me know if you find any bugs Thanks -Nick
filter field have to be indexed?
i means define it in schema.xml,,, -- regards jl
Re: i wanna change response type to PHP serialize
i just cp it to src\java\org\apache\solr\request and ant dist...i think maybe my method is wrong. same error infromation... java.lang.NoClassDefFoundError: org/apache/solr/search/ScorePriorityQueue at org.apache.solr.search.SolrIndexSearcher.getDocListNC(SolrIndexSearcher.java:886) at org.apache.solr.search.SolrIndexSearcher.getDocListC(SolrIndexSearcher.java:805) at org.apache.solr.search.SolrIndexSearcher.getDocList(SolrIndexSearcher.java:698) at org.apache.solr.request.StandardRequestHandler.handleRequestBody(StandardRequestHandler.java:122) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:77) at org.apache.solr.core.SolrCore.execute(SolrCore.java:658) at org.apache.solr.servlet.SolrServlet.doGet(SolrServlet.java:66) at javax.servlet.http.HttpServlet.service(HttpServlet.java:690) at javax.servlet.http.HttpServlet.service(HttpServlet.java:803) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:290) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:185) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:228) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:175) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:104) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:216) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:844) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:634) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:445) at java.lang.Thread.run(Unknown Source) 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I have some good news :o) https://issues.apache.org/jira/browse/SOLR-275 Please let me know if you find any bugs Thanks -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: I think it simple to u. so i wait for ur good news. 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I am also quite interested in getting a serialized PHP array response writer, after some investigation it doesn't seem as difficult as I first thought, I will have a try at implementing this when I get some time, the format of the array would probably end up being the same as if you were to use json_decode. -Nick On 6/26/07, Tristan Vittorio [EMAIL PROTECTED] wrote: I am also looking for a PHP response writer that outputs serialized PHP code, I've taken a look at SOLR-196http://issues.apache.org/jira/browse/SOLR-196but it's not quite what I need. Unless anyone else has started implementing one, I will see if I can modify Paul's patch to do the task. cheers, Tristan On 6/25/07, Michael Kimsal [EMAIL PROTECTED] wrote: My experience tells me that serializing/unserializing tends to be faster than eval(), especially on larger arrays. I've looked at the code, trying to put together a PHP serialized array output type, but haven't had success doing it yet. If I do, it'll be contributed back, but until someone does that version, the way Nick suggests is probably the only option. On 6/25/07, Nick Jenkin [EMAIL PROTECTED] wrote: Hi James I think you would be better of outputting an PHP array, and running eval() over it, the PHP serialize format is quite complicated. On that note, you might be interested in: http://issues.apache.org/jira/browse/SOLR-196 -Nick On 6/25/07, James liu [EMAIL PROTECTED] wrote: which files i should change from source? and if i change ok. how to compile? just ant dist? -- regards jl -- Michael Kimsal http://webdevradio.com -- regards jl -- regards jl
Re: snapshooter no go
first try it? which system u use? if u use freebsd, just give up trying. it not fit for freebsd. 2007/6/27, Otis Gospodnetic [EMAIL PROTECTED]: Hi, Here is a puzzling one. I can't get Solr to invoke snaphooter properly. Solr claims my snapshooter is not where I said it is: SEVERE: java.io.IOException: Cannot run program snapshooter (in directory solr/bin): java.io.IOException: error=2, No such file or directory So I added some logging to RunExecutableListener: INFO: CMD: snapshooter INFO: ENV: MYVAR=val1 INFO: DIR: /foobar/jetty-6.1.3/solr/bin This looks correct, and I can see snapshooter is there: $ cd /foo/jetty-6.1.3 $ ls -al solr/bin/snapshooter -rwxr-xr-x 1 otis otis 2624 Jun 20 02:53 solr/bin/snapshooter So snapshooter *does* exist under /foobar/jetty-6.1.3/solr/bin and is executable by everyone. I've tried moving solr home dir outside the jetty dir, but that made no difference. I tried specifying the absolute path to the dir parameter in the config for snapshooter in solrconfig.xml, again no luck. Huh? Is this some new bug? Has anyone seen this before? I'm using the HEAD. Thanks, Otis -- regards jl
Re: Does sort field need have to be 'store'?
2007/6/27, Mike Klaas [EMAIL PROTECTED]: On 25-Jun-07, at 10:53 PM, James liu wrote: [quote]how can i use index all with ram and how to config which ram i should use?[/quote] Your os will automatically load the most frequently-used parts of the index in ram. If your total ram is at least as large as the total JVM heap size plus the on-disk footprint of the index, it will eventually reside completely in memory (assuming no other processes are using ram). I don't understand your second question. -MIke thks mike,,,the second question is how to calc the ram size i should have when i wanna index reside completely in RAM. -- regards jl
Re: i wanna change response type to PHP serialize
Yes, i use 1.2my compile method: download solr 1.2 and modify file by your patch. use `ant dist` to compile it. no error show. i can see its admin gui, but i wanna try search, it will show me the error information, Is my compile method right? if not, show me how to compile it. it seems very strange ,only me fail? anyone have same question? if free, maybe u zip your solr to me by mail...and i try it again. 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: Interesting, what version of solr are you using, I tested on 1.2. -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: i just cp it to src\java\org\apache\solr\request and ant dist...i think maybe my method is wrong. same error infromation... java.lang.NoClassDefFoundError: org/apache/solr/search/ScorePriorityQueue at org.apache.solr.search.SolrIndexSearcher.getDocListNC( SolrIndexSearcher.java:886) at org.apache.solr.search.SolrIndexSearcher.getDocListC( SolrIndexSearcher.java:805) at org.apache.solr.search.SolrIndexSearcher.getDocList( SolrIndexSearcher.java:698) at org.apache.solr.request.StandardRequestHandler.handleRequestBody( StandardRequestHandler.java:122) at org.apache.solr.handler.RequestHandlerBase.handleRequest( RequestHandlerBase.java:77) at org.apache.solr.core.SolrCore.execute(SolrCore.java:658) at org.apache.solr.servlet.SolrServlet.doGet(SolrServlet.java :66) at javax.servlet.http.HttpServlet.service(HttpServlet.java:690) at javax.servlet.http.HttpServlet.service(HttpServlet.java:803) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter( ApplicationFilterChain.java:290) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.solr.servlet.SolrDispatchFilter.doFilter( SolrDispatchFilter.java:185) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter( ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke( StandardWrapperValve.java:228) at org.apache.catalina.core.StandardContextValve.invoke( StandardContextValve.java:175) at org.apache.catalina.core.StandardHostValve.invoke( StandardHostValve.java:128) at org.apache.catalina.valves.ErrorReportValve.invoke( ErrorReportValve.java:104) at org.apache.catalina.core.StandardEngineValve.invoke( StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service( CoyoteAdapter.java:216) at org.apache.coyote.http11.Http11Processor.process( Http11Processor.java:844) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process( Http11Protocol.java:634) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run( JIoEndpoint.java:445) at java.lang.Thread.run(Unknown Source) 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I have some good news :o) https://issues.apache.org/jira/browse/SOLR-275 Please let me know if you find any bugs Thanks -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: I think it simple to u. so i wait for ur good news. 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I am also quite interested in getting a serialized PHP array response writer, after some investigation it doesn't seem as difficult as I first thought, I will have a try at implementing this when I get some time, the format of the array would probably end up being the same as if you were to use json_decode. -Nick On 6/26/07, Tristan Vittorio [EMAIL PROTECTED] wrote: I am also looking for a PHP response writer that outputs serialized PHP code, I've taken a look at SOLR-196http://issues.apache.org/jira/browse/SOLR-196but it's not quite what I need. Unless anyone else has started implementing one, I will see if I can modify Paul's patch to do the task. cheers, Tristan On 6/25/07, Michael Kimsal [EMAIL PROTECTED] wrote: My experience tells me that serializing/unserializing tends to be faster than eval(), especially on larger arrays. I've looked at the code, trying to put together a PHP serialized array output type, but haven't had success doing it yet. If I do, it'll be contributed back, but until someone does that version, the way Nick suggests is probably the only option. On 6/25/07, Nick Jenkin [EMAIL PROTECTED] wrote: Hi James I think you would be better of outputting an PHP array, and running eval() over it, the PHP serialize format is quite complicated. On that note, you might be interested in: http://issues.apache.org/jira/browse/SOLR-196 -Nick On 6/25/07, James liu [EMAIL PROTECTED
Re: I'm using PHP curl post xml command to Solr,Is it the only way to post data?
how about its performance? 2007/6/26, Kijiji Xu, Ping [EMAIL PROTECTED]: I had solved this problem,below is my POST code,I used HTTP_Request of PEAR,it's so simple.thank you all very much .FYI; private function doPost($url,$postData){ $req = new HTTP_Request($url,array( 'method' = HTTP_REQUEST_METHOD_POST, 'http' = HTTP_REQUEST_HTTP_VER_1_1)); $req-addHeader('Content-Type','text/xml; charset=utf-8'); $req-addHeader('Content-Length',strlen($postData)); $req-addRawPostData($postData,true); $req-sendRequest(); $response1 = $req-getResponseBody(); echo $response1.\n; return $response1; } -Original Message- From: Tristan Vittorio [mailto:[EMAIL PROTECTED] Sent: 2007年6月25日 21:36 To: solr-user@lucene.apache.org Subject: Re: I'm using PHP curl post xml command to Solr,Is it the only way to post data? Rather than re-inventing the wheel with fsockopen, why don't you use an existing HTTP client implementation written in PHP: http://pear.php.net/package/HTTP_Client http://framework.zend.com/manual/en/zend.http.html http://scripts.incutio.com/httpclient/ http://www.phpclasses.org/browse/package/576.html They should all be able to perform the HTTP POST requests that you require and will be a lot easier to use that writing raw HTTP protocol through a socket! regards, Tristan On 6/25/07, Kijiji Xu, Ping [EMAIL PROTECTED] wrote: In our product environment, there haven't curl and pecl lib,so... If the fsockopen could solve post xml data,I would like to use it.And I tried it, but I failed. -Original Message- From: Michael Kimsal [mailto:[EMAIL PROTECTED] Sent: 2007年6月25日 20:16 To: solr-user@lucene.apache.org Subject: Re: I'm using PHP curl post xml command to Solr,Is it the only way to post data? Using PHP5 (5.1 or higher I think) http://us.php.net/manual/en/function.http-post-fields.php is available. From the example on that page: $fields = array( 'name' = 'mike', 'pass' = 'passwordt' ); $response = http_post_fields(http://www.example.com/;, $fields); Looks pretty simple, but I haven't tried it yet. On 6/25/07, Kijiji Xu, Ping [EMAIL PROTECTED] wrote: What about fsockopen, Or any other simple method? Thanks -- Regards Xp from china -- Michael Kimsal http://webdevradio.com -- regards jl
Re: i wanna change response type to PHP serialize
i use tomcat ,, send ur solr version to me...i try it again.. 2007/6/27, Nick Jenkin [EMAIL PROTECTED]: If you are using the example provided in 1.2 (using jetty) you need to use ant example rather than ant dist -Nick On 6/27/07, James liu [EMAIL PROTECTED] wrote: Yes, i use 1.2my compile method: download solr 1.2 and modify file by your patch. use `ant dist` to compile it. no error show. i can see its admin gui, but i wanna try search, it will show me the error information, Is my compile method right? if not, show me how to compile it. it seems very strange ,only me fail? anyone have same question? if free, maybe u zip your solr to me by mail...and i try it again. 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: Interesting, what version of solr are you using, I tested on 1.2. -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: i just cp it to src\java\org\apache\solr\request and ant dist...i think maybe my method is wrong. same error infromation... java.lang.NoClassDefFoundError: org/apache/solr/search/ScorePriorityQueue at org.apache.solr.search.SolrIndexSearcher.getDocListNC( SolrIndexSearcher.java:886) at org.apache.solr.search.SolrIndexSearcher.getDocListC( SolrIndexSearcher.java:805) at org.apache.solr.search.SolrIndexSearcher.getDocList( SolrIndexSearcher.java:698) at org.apache.solr.request.StandardRequestHandler.handleRequestBody( StandardRequestHandler.java:122) at org.apache.solr.handler.RequestHandlerBase.handleRequest( RequestHandlerBase.java:77) at org.apache.solr.core.SolrCore.execute(SolrCore.java:658) at org.apache.solr.servlet.SolrServlet.doGet( SolrServlet.java :66) at javax.servlet.http.HttpServlet.service(HttpServlet.java :690) at javax.servlet.http.HttpServlet.service(HttpServlet.java :803) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter( ApplicationFilterChain.java:290) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.solr.servlet.SolrDispatchFilter.doFilter( SolrDispatchFilter.java:185) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter( ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke( StandardWrapperValve.java:228) at org.apache.catalina.core.StandardContextValve.invoke( StandardContextValve.java:175) at org.apache.catalina.core.StandardHostValve.invoke( StandardHostValve.java:128) at org.apache.catalina.valves.ErrorReportValve.invoke( ErrorReportValve.java:104) at org.apache.catalina.core.StandardEngineValve.invoke( StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service( CoyoteAdapter.java:216) at org.apache.coyote.http11.Http11Processor.process( Http11Processor.java:844) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process( Http11Protocol.java:634) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run( JIoEndpoint.java:445) at java.lang.Thread.run(Unknown Source) 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I have some good news :o) https://issues.apache.org/jira/browse/SOLR-275 Please let me know if you find any bugs Thanks -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: I think it simple to u. so i wait for ur good news. 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I am also quite interested in getting a serialized PHP array response writer, after some investigation it doesn't seem as difficult as I first thought, I will have a try at implementing this when I get some time, the format of the array would probably end up being the same as if you were to use json_decode. -Nick On 6/26/07, Tristan Vittorio [EMAIL PROTECTED] wrote: I am also looking for a PHP response writer that outputs serialized PHP code, I've taken a look at SOLR-196http://issues.apache.org/jira/browse/SOLR-196but it's not quite what I need. Unless anyone else has started implementing one, I will see if I can modify Paul's patch to do the task. cheers, Tristan On 6/25/07, Michael Kimsal [EMAIL PROTECTED] wrote: My experience tells me that serializing/unserializing tends to be faster than eval(), especially on larger arrays. I've looked at the code, trying to put together a PHP serialized array output type, but haven't had success doing it yet. If I do, it'll be contributed back
Re: i wanna change response type to PHP serialize
ok,,thks nick,,,i just forget replace jar file.. wait a minute i will test speed... 2007/6/27, Nick Jenkin [EMAIL PROTECTED]: http://nickjenkin.com/misc/apache-solr-1.2.0-php-serialize.tar.gz Try that -Nick On 6/27/07, James liu [EMAIL PROTECTED] wrote: i use tomcat ,, send ur solr version to me...i try it again.. 2007/6/27, Nick Jenkin [EMAIL PROTECTED]: If you are using the example provided in 1.2 (using jetty) you need to use ant example rather than ant dist -Nick On 6/27/07, James liu [EMAIL PROTECTED] wrote: Yes, i use 1.2my compile method: download solr 1.2 and modify file by your patch. use `ant dist` to compile it. no error show. i can see its admin gui, but i wanna try search, it will show me the error information, Is my compile method right? if not, show me how to compile it. it seems very strange ,only me fail? anyone have same question? if free, maybe u zip your solr to me by mail...and i try it again. 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: Interesting, what version of solr are you using, I tested on 1.2. -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: i just cp it to src\java\org\apache\solr\request and ant dist...i think maybe my method is wrong. same error infromation... java.lang.NoClassDefFoundError: org/apache/solr/search/ScorePriorityQueue at org.apache.solr.search.SolrIndexSearcher.getDocListNC ( SolrIndexSearcher.java:886) at org.apache.solr.search.SolrIndexSearcher.getDocListC( SolrIndexSearcher.java:805) at org.apache.solr.search.SolrIndexSearcher.getDocList( SolrIndexSearcher.java:698) at org.apache.solr.request.StandardRequestHandler.handleRequestBody( StandardRequestHandler.java:122) at org.apache.solr.handler.RequestHandlerBase.handleRequest( RequestHandlerBase.java:77) at org.apache.solr.core.SolrCore.execute(SolrCore.java :658) at org.apache.solr.servlet.SolrServlet.doGet( SolrServlet.java :66) at javax.servlet.http.HttpServlet.service( HttpServlet.java :690) at javax.servlet.http.HttpServlet.service( HttpServlet.java :803) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter( ApplicationFilterChain.java:290) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.solr.servlet.SolrDispatchFilter.doFilter( SolrDispatchFilter.java:185) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter( ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke( StandardWrapperValve.java:228) at org.apache.catalina.core.StandardContextValve.invoke( StandardContextValve.java:175) at org.apache.catalina.core.StandardHostValve.invoke( StandardHostValve.java:128) at org.apache.catalina.valves.ErrorReportValve.invoke( ErrorReportValve.java:104) at org.apache.catalina.core.StandardEngineValve.invoke( StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service( CoyoteAdapter.java:216) at org.apache.coyote.http11.Http11Processor.process( Http11Processor.java:844) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process( Http11Protocol.java:634) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run( JIoEndpoint.java:445) at java.lang.Thread.run(Unknown Source) 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I have some good news :o) https://issues.apache.org/jira/browse/SOLR-275 Please let me know if you find any bugs Thanks -Nick On 6/26/07, James liu [EMAIL PROTECTED] wrote: I think it simple to u. so i wait for ur good news. 2007/6/26, Nick Jenkin [EMAIL PROTECTED]: I am also quite interested in getting a serialized PHP array response writer, after some investigation it doesn't seem as difficult as I first thought, I will have a try at implementing this when I get some time, the format of the array would probably end up being the same as if you were to use json_decode. -Nick On 6/26/07, Tristan Vittorio [EMAIL PROTECTED] wrote: I am also looking for a PHP response writer that outputs serialized PHP code, I've taken a look at SOLR-196http://issues.apache.org/jira/browse/SOLR-196 but it's not quite what I need. Unless anyone else has started
Re: i wanna change response type to PHP serialize
i use stupid way just to create file by your patch. and compile is ok when i test it and start tomcat,,,i can see solr admin interface. but when i search something.. it show me 2007-6-25 15:15:37 org.apache.solr.core.SolrException log 严重: java.lang.NoClassDefFoundError: org/apache/solr/search/ScorePriorityQueue at org.apache.solr.search.SolrIndexSearcher.getDocListNC( SolrIndexSearcher.java:886) at org.apache.solr.search.SolrIndexSearcher.getDocListC( SolrIndexSearcher.java:805) at org.apache.solr.search.SolrIndexSearcher.getDocList( SolrIndexSearcher.java:698) at org.apache.solr.request.StandardRequestHandler.handleRequestBody( StandardRequestHandler.java:122) at org.apache.solr.handler.RequestHandlerBase.handleRequest( RequestHandlerBase.java:77) at org.apache.solr.core.SolrCore.execute(SolrCore.java:658) at org.apache.solr.servlet.SolrServlet.doGet(SolrServlet.java:66) at javax.servlet.http.HttpServlet.service(HttpServlet.java:690) at javax.servlet.http.HttpServlet.service(HttpServlet.java:803) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter( ApplicationFilterChain.java:290) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.solr.servlet.SolrDispatchFilter.doFilter( SolrDispatchFilter.java:185) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter( ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter( ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke( StandardWrapperValve.java:228) at org.apache.catalina.core.StandardContextValve.invoke( StandardContextValve.java:175) at org.apache.catalina.core.StandardHostValve.invoke( StandardHostValve.java:128) at org.apache.catalina.valves.ErrorReportValve.invoke( ErrorReportValve.java:104) at org.apache.catalina.core.StandardEngineValve.invoke( StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service( CoyoteAdapter.java:216) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java :844) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process( Http11Protocol.java:634) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java :445) at java.lang.Thread.run(Unknown Source) i find ScorePriorityQueue in SolrIndexSearcher, and i no change it. 2007/6/25, James liu [EMAIL PROTECTED]: I means how to add it to my solr(1.2 production) 2007/6/25, James liu [EMAIL PROTECTED]: aha,,it seems good, how can i fix it with my solr, i don't know how do with it 2007/6/25, Nick Jenkin [EMAIL PROTECTED]: Hi James I think you would be better of outputting an PHP array, and running eval() over it, the PHP serialize format is quite complicated. On that note, you might be interested in: http://issues.apache.org/jira/browse/SOLR-196 -Nick On 6/25/07, James liu [EMAIL PROTECTED] wrote: which files i should change from source? and if i change ok. how to compile? just ant dist? -- regards jl -- regards jl -- regards jl -- regards jl
Re: Does sort field need have to be 'store'?
thks Yonik,and [quote]how can i use index all with ram and how to config which ram i should use?[/quote]
i wanna change response type to PHP serialize
which files i should change from source? and if i change ok. how to compile? just ant dist? -- regards jl
Re: i wanna change response type to PHP serialize
aha,,it seems good, how can i fix it with my solr, i don't know how do with it 2007/6/25, Nick Jenkin [EMAIL PROTECTED]: Hi James I think you would be better of outputting an PHP array, and running eval() over it, the PHP serialize format is quite complicated. On that note, you might be interested in: http://issues.apache.org/jira/browse/SOLR-196 -Nick On 6/25/07, James liu [EMAIL PROTECTED] wrote: which files i should change from source? and if i change ok. how to compile? just ant dist? -- regards jl -- regards jl
Re: i wanna change response type to PHP serialize
I means how to add it to my solr(1.2 production) 2007/6/25, James liu [EMAIL PROTECTED]: aha,,it seems good, how can i fix it with my solr, i don't know how do with it 2007/6/25, Nick Jenkin [EMAIL PROTECTED]: Hi James I think you would be better of outputting an PHP array, and running eval() over it, the PHP serialize format is quite complicated. On that note, you might be interested in: http://issues.apache.org/jira/browse/SOLR-196 -Nick On 6/25/07, James liu [EMAIL PROTECTED] wrote: which files i should change from source? and if i change ok. how to compile? just ant dist? -- regards jl -- regards jl -- regards jl
Re: commit script with solr 1.2 response format
aha,,same question i found few days ago. i m sorry to forget submit it. 2007/6/22, Yonik Seeley [EMAIL PROTECTED]: On 6/21/07, Ryan McKinley [EMAIL PROTECTED] wrote: I just started running the scripts and The commit script seems to run fine, but it says there was an error. I looked into it, and the scripts expect 1.1 style response: result status=0/result 1.2 /update returns: ?xml version=1.0 encoding=UTF-8? response lst name=responseHeader int name=status0/int int name=QTime44/int /lst /response I guess we should look for 'status=0' ? Or, if you get a response code of 200, it's a success unless you see status=nonzero -Yonik -- regards jl
Re: Slave/Master swap
If just one master or one slave server fail, i think u maybe can use master index server. shell controlled by program is easy for me. i use php and shell_exec. 2007/6/21, Otis Gospodnetic [EMAIL PROTECTED]: Right, that SAN con 2 Masters sounds good. Lucky you with your lonely Master! Where I work hw failures are pretty common. Otis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Simpy -- http://www.simpy.com/ - Tag - Search - Share - Original Message From: Chris Hostetter [EMAIL PROTECTED] To: solr-user@lucene.apache.org Sent: Wednesday, June 20, 2007 11:43:02 PM Subject: Re: Slave/Master swap : The more expensive solution might be to have Solr instances run on top : of a SAN and then one could really have multiple Master instances, one : in stand-by mode and ready to be started as the new Master if the i *believe* that if you have two solr isntances pointed at the same physical data directory (SAN or otherwise) but you only send update/commit commands to one, they won't interfere with eachother. so concievable you can have both masters up and running and your failover approach if the primary goes down is just to start sending updates to the secondary. you'll loose any unflushed changes that hte primary had in memory, but those are lost anyway. don't trust me on that though, test it out yourself. : curiosity, how does CNet handle Master redundancy? I don't know how much i'm allowed to talk about our processes and systems for redundency, disastery recovery, fallover, etc... but i don't think i'll upset anyone if i tell you: as far as i know, we've never needed to take advantage of them with a solr master. ie: we've never had a solr master crash so hard we had to bring up another one in it's place ... knock on wood. (that probably has more to do with having good hardware then anything else though). (and no, i honestly don't know what hardware we use ... i don't bother paying attention, i let hte hardware guys worry about that) -Hoss -- regards jl
Re: Multiple doc types in schema
I see SOLR-215 from this mail. Does it now really support multi index and search it will return merged data? for example: i wanna search: aaa, and i have index1, index2, index3, index4it should return the result from index1,index2,index3, index4 and merge result by score, datetime, or other thing. Does it support NFS and how its performance? 2007/6/21, Otis Gospodnetic [EMAIL PROTECTED]: This sounds like a potentially good use-case for SOLR-215! See https://issues.apache.org/jira/browse/SOLR-215 Otis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Simpy -- http://www.simpy.com/ - Tag - Search - Share - Original Message From: Chris Hostetter [EMAIL PROTECTED] To: solr-user@lucene.apache.org; Jack L [EMAIL PROTECTED] Sent: Wednesday, June 6, 2007 6:58:10 AM Subject: Re: Multiple doc types in schema : This is based on my understanding that solr/lucene does not : have the concept of document type. It only sees fields. : : Is my understanding correct? it is. : It seems a bit unclean to mix fields of all document types : in the same schema though. Or, is there a way to allow multiple : document types in the schema, and specify what type to use : when indexing and searching? it's really just an issue of semantics ... the schema.xml is where you list all of the fields you need in your index, any notion of doctype is entire artificial ... you could group all of the fields relating to doctypeA in one section of the schema.xml, then have a big !-- ##...## -- line and then list the fields in doctypeB, etc... but wat if there are fields you use in both doctypes ? .. how much you mix them is entirely up to you. -Hoss -- regards jl
delete changed?
solr:1.2 curl http://192.168.7.6:8080/solr0/update --data-binary 'deletequerynodeid:20/query/delete' i remember it is ok when i use solr 1.1 does it change? it show me: HTTP Status 400 - missing content stream -- *type* Status report *message* *missing content stream* *description* *The request sent by the client was syntactically incorrect (missing content stream).* -- regards jl
it is strange.
It seems strange when i refresh same url search. time will change...sometime use *0.01021409034729 s, *sometime use * 0.0080091953277588 s. *sometime use *0.024219989776611. It change too big. * Only i use it and less search, so i think memory not all use. why time changed very big, and i think it should change in a small range. -- regards jl
Re: it is strange.
ok, i find it only happen in win. 2007/6/19, James liu [EMAIL PROTECTED]: It seems strange when i refresh same url search. time will change...sometime use *0.01021409034729 s, *sometime use * 0.0080091953277588 s. *sometime use *0.024219989776611 . It change too big. * Only i use it and less search, so i think memory not all use. why time changed very big, and i think it should change in a small range. -- regards jl -- regards jl
faceted query result
for example. i wanna show keyword: a and facet sid: 2 my url: http://localhost:8080/solr1/select?q=a+sid:2start=0rows=10fl=*wt=json but it show me count bigger than facetnum. i read http://lucene.apache.org/java/docs/queryparsersyntax.html and try server way , all not effect. maybe someone know it. -- regards jl
Re: faceted query result
thks. 2007/6/17, Yonik Seeley [EMAIL PROTECTED]: On 6/16/07, James liu [EMAIL PROTECTED] wrote: i wanna show keyword: a and facet sid: 2 my url: http://localhost:8080/solr1/select?q=a+sid:2start=0rows=10fl=*wt=json but it show me count bigger than facetnum. '+' in a URL is like a space, so your query is the same as default_field:a OR sid:2 instead of default_field:a AND sid:2 This should work: q=a+%2Bsid:2 or preferably use a filter query q=afq=sid:2 -Yonik -- regards jl
who use time?
i write script to get run time to sure how to performance. i find very intresting thing that i query 2 solr box to get data and solr response show me qtime all zero. but i find multi get data script use time is 0.046674966812134(it will change) solr box in my pc. and index data is very small. so i don't know why it use much time like 0.046674966812134. -- regards jl
Re: problems getting data into solr index
is it ok? 2007/6/14, vanderkerkoff [EMAIL PROTECTED]: Hi Yonik Here's the output from netcat POST /solr/update HTTP/1.1 Host: localhost:8983 Accept-Encoding: identity Content-Length: 83 Content-Type: text/xml; charset=utf-8 that looks Ok to me, but I am a bit twp you see. :-) Yonik Seeley wrote: On 6/13/07, vanderkerkoff [EMAIL PROTECTED] wrote: I'm running solr1.2 and Jetty, I'm having problems looping through a mysql database with python and putting the data into the solr index. Here's the error UnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 369: ordinal not in range(128) There are two issues... what char encoding you tell solr to use, via Content-type in the HTTP headers (solr defaults to UTF-8), and then if what you send matches that coding. If you can get the complete message (including HTTP headers) that is being sent to Solr, that would help people debug the problem. One easy way is to use netcat to pretend to be solr: 1) shut down solr 2) start up netcat on solr's port nc -l -p 8983 3) send your update message from the client as you normally would -Yonik -- View this message in context: http://www.nabble.com/problems-getting-data-into-solr-index-tf3915542.html#a6020 Sent from the Solr - User mailing list archive at Nabble.com. -- regards jl
Re: who use time?
2007/6/14, Yonik Seeley [EMAIL PROTECTED]: On 6/14/07, James liu [EMAIL PROTECTED] wrote: i write script to get run time to sure how to performance. i find very intresting thing that i query 2 solr box to get data and solr response show me qtime all zero. but i find multi get data script use time is 0.046674966812134(it will change) If you are timing the complete script there is startup time to take into account. If you are only timing the request/response to solr, then that is a bit slow considering the query time itself is less than a millisecond. That does not include document retrieval and response writing. I just timing my script to get data from 2 solr boxes, not complete script. It just query two box and return id,score .rows=10. response type use json. and i see their qtime all zero. How many documents are you retrieving? one: numDocs : 1 maxDoc : 23000 the other: numDocs : 9000 maxDoc : 9000 3-4k per doc If you re-execute the same exact query again, is it still slower? It will be quick. time will be 0.0043279ms -Yonik -- regards jl
Re: Where are the log files...
if u use jetty, u should see jetty's log. if u use tomcat, u should see tomcat's log. solr is only a program that run with container. 2007/6/15, Ryan McKinley [EMAIL PROTECTED]: what version of solr/container are you running? this sounds similar to what people running solr 1.1 with the jetty include in that example... Jack L wrote: It happened twice in the past few days that the solr instance stopped responding (the admin page does not load) while the process was still running. I'd like to find out what's causing this. I notice that I can change logger level from admin page but I didn't figure out where the log files are. Search on solr wiki and internet didn't help... -- regards jl
solr+hadoop = next solr
anyone agree? Next solr's development 's plan is? anyone know? -- regards jl
Re: solr+hadoop = next solr
2007/6/7, Yonik Seeley [EMAIL PROTECTED]: On 6/6/07, James liu [EMAIL PROTECTED] wrote: anyone agree? No ;-) At least not if you mean using map-reduce for queries. When I started looking at distributed search, I immediately went and read the map-reduce paper (easier concept than it first appeared), and realized it's really more for the indexing side of things (big batch jobs, making data from data, etc). Nutch uses map reduce for crawling/indexing, but not for querying. Yes, nutch use map reduce only for crawling/indexing, not for querying. http://www.nabble.com/something-i-think-important-and-should-be-added-tf3813838.html#a10796136 map-reduce just for indexing to decrease Master solr query *instance *index size and increase query speed. It will cost many time to index and merge but it will increase query accuracy. index and data not in same box. so we just only sure master query server hardware is powerful and slave query server hardware is not very important. Master index server should support multi index. If solr support it. I think user who use solr will quick setup their search. It just my thought. how do u think, yonik,,,and how do u think next solr? -Yonik -- regards jl
welcome solr1.2
thks Solr Committers -- regards jl
can't start solr by tomcat
solr 1.3dev 2007-06-04(svn) tomcat log show me error information: solr 1.3dev 2007-06-04 org.apache.solr.core.SolrException: Unknown fieldtype 'string' i find it only use in shema.xml field name=id type=string indexed=true stored=true required=true / i modifiy it and now start is ok field name=id type=integer stored=true / property required means? i not find it in comment. -- regards jl
Re: can't start solr by tomcat
thks,ryan, i find required in changes.txt 2007/6/4, Ryan McKinley [EMAIL PROTECTED]: i modifiy it and now start is ok field name=id type=integer stored=true / property required means? i not find it in comment. required means that the field *must* be specified when you add it to the index. If it isn't there, you will get an error. If you upgrade or work from trunk, make sure to read CHANGES.txt carefully -- when 1.2 is officially released (soon I swear!) this will be made very clear. ryan -- regards jl
Re: facet should add facet.analyzer
2007/5/29, Chris Hostetter [EMAIL PROTECTED]: : facet.analyzer is true, do analyze, if false don't analyze. : What if Solr doesn't have access to the unindexed version? My : suggestion would be to copyField into an unanalyzed version, and : facet on that. me too. yeah, i'm not even sure that i understand the orriginal suggestion clearly (i'd need a more concrete example) but analysis doesn't just happen at query i know it. (or facet) time) ... there's also the indexing aspect of it. in general, options which disable query time analysis don't really make sense because no query time option can undo the fact that index time analysis has already taken place. it seems that only map facet value by solr or by web(if use web language) to avoid facet value be analyzed(english value have space, chinese value will be analyzed.). -Hoss -- regards jl
facet should add facet.analyzer
facet.analyzer is true, do analyze, if false don't analyze. why i say that, Chinese word not use space to split, so if analyzed, it will change. now i will use map to fix it before no facet.analyzer. -- regards jl
Re: index problem with write lock
i find it always happen when index have been doing for a while. for example, it will happen after starting index 1 hour - 2hours. 2007/5/24, James liu [EMAIL PROTECTED]: i find one interesting thing. when i index data with 45 solr boxs.(data have 1700w, freebsd6, java: diablo-1.5.0_07-b01, tomcat6), write lock will happen in the procedure. Reindex with solr box which have problem with write block. it show me well. it happen serveral times, so i wanna know why it happen, it should be ok in theory. not every solr box will have this problem. 45 solr boxs have same config and empty index.(i use copy, just rename their directory name) anyone have same problem and know why? -- regards jl -- regards jl
Re: index problem with write lock
2007/5/25, Chris Hostetter [EMAIL PROTECTED]: : when i index data with 45 solr boxs.(data have 1700w, freebsd6, java: : diablo-1.5.0_07-b01, tomcat6), write lock will happen in the procedure. 1) bug reports about errors are nearly useless without a real error message including a stack trace. 2) what do you mean you index data with 45 solr boxs .. are you running 45 seperate instances of solr and idnexing on all of them indepndently? if so why doees the number matter? .. it sounds like you are describing a problem you would habe after a while even if there was only 1 solr server right? 45 solr boxs means 45 seperate instances of solr . The number why i say because my code use for statement to index . like this: for($i=0; $i45; $i++){ doIndex($i); } 1700w data divided into 45, and sent to 45 solr instances. typically when i see problems with write locks it's because Solr crashed (usually from an OOM) and then the container restarted it but the stale write lock was still on disk ... have you checked your logs for other previous exceptions? i know how to fix it. but i just don't know why it happen. this solr error information: Exception during commit/optimize:java.io.IOException: Lock obtain timed out: SimpleFSLock@/usr/solrapp/solr21/data/index/write.lock -Hoss Thks hoss. -- regards jl
something i think important and should be added
multi layer: now solr's procedure: user query - solr instance - show results i think it maybe simple to some application maybe this procedure fit for: user query - Master solr query instance - single solr query instance - show results master solr query instance: it can define some global field, like score, time, userid which is from every single solr query instance. if so, we can do sort by score, sort by time, filter by userid and so on. To solr user, we not concert about how to do multi query in parrel and merge data . We just care how to define global field, our busness, search ui. test: why we test, we just wanna optimize. how to optimize, which data range i should do something like 'add memory','add solr instance', 'lower filter cache', 'increase query cache', 'less memory so lower autowarm'. now i only depend on my experience to adjust cache, memory, docs num and i do'nt which best i should. so if solr have tools to test or moniter(i read its discuss in pre mail). i read article from wiki ,but i not find something about data range. so i think public tools which test and get data, optimize solr with that data. Public moniter which can tell me real time data change maybe these all solr tomorrow plan, if u have experience with solr, let's talk about it . i like solr. -- regards jl
Re: Difficulty posting unicode to solr index
how do u sure ur file is encoded by utf-8? 2007/5/24, Ethan Gruber [EMAIL PROTECTED]: Hi, I am attempting to post some unicode XML documents to my solr index. They are encoded in UTF-8. When I attempt to query from the solr admin page, I'm basically getting gibberish garbage text in return. I decided to try a file that I know is supposed to work, which is the utf8-example.xml found in the exampledocs folder. This also did not return proper unicode results. None of my other coworkers have run into this problem, but I believe there is one difference between their system and my system which could account for the error. They're using Macs and thus posting with post.sh, and I am running Windows and posting with a post.jar file. Could post.jar not support unicode? Has anyone run into this problem before? Thanks, Ethan -- regards jl
Re: List of highlighted terms from search query
2007/5/23, solruser [EMAIL PROTECTED]: Hi James, I have already had tried out options for highlighting but unfortunately since solr highligted results does not include special characters special characters? maybe u should print some . in the results, thats the reason I am looking for a way to get list of highlighted terms that I can use to highlight the terms. I assume this might be issue which must been picked in the mailing list earlier too but I do not get relevant solution. Another thing I want to get the list of terms in hope to get better alternative. hope I make myself clear here Thanks -Amit James liu wrote: first u try enable highlighting( http://wiki.apache.org/solr/HighlightingParameters) and u try solr admin gui to see its output and u will find what u wanna. 2007/5/23, solruser [EMAIL PROTECTED]: Hi, I am wondering can we get the list of all highlighted terms from the search query. If yes can someone suggest query options? Thanks -- View this message in context: http://www.nabble.com/List-of-highlighted-terms-from-search-query-tf3801344.html#a10755381 Sent from the Solr - User mailing list archive at Nabble.com. -- regards jl -- View this message in context: http://www.nabble.com/List-of-highlighted-terms-from-search-query-tf3801344.html#a10757932 Sent from the Solr - User mailing list archive at Nabble.com. -- regards jl
Re: PriceJunkie.com using solr!
ahait is wonderful. 2007/5/24, Mike Austin [EMAIL PROTECTED]: Just one. -Original Message- From: James liu [mailto:[EMAIL PROTECTED] Sent: Wednesday, May 16, 2007 10:30 PM To: solr-user@lucene.apache.org Subject: Re: PriceJunkie.com using solr! how many solr instance? 2007/5/17, Yonik Seeley [EMAIL PROTECTED]: Congrats, very nice job! It's fast too. -Yonik On 5/16/07, Mike Austin [EMAIL PROTECTED] wrote: I just wanted to say thanks to everyone for the creation of solr. I've been using it for a while now and I have recently brought one of my side projects online. I have several other projects that will be using solr for it's search and facets. Please check out www.pricejunkie.com and let us know what you think.. You can give feedback and/or sign up on the mailing list for future updates. The site is very basic right now and many new and useful features plus merchants and product categories will be coming soon! I thought it would be a good idea to at least have a few people use it to get some feedback early and often. Some of the nice things behind the scenes that we did with solr: - created custom request handlers that have category to facet to attribute caching built in - category to facet management - ability to manage facet groups (attributes within a set facet) and assign them to categories - ability to create any category structure and share facet groups - facet inheritance for any category (a facet group can be defined on a parent category and pushed down to all children) - ability to create sub-categories as facets instead of normal sub categories - simple xml configuration for the final outputted category configuration file I'm sure there are more cool things but that is all for now. Join the mailing list to see more improvements in the future. Also.. how do I get added to the Using Solr wiki page? Thanks, Mike Austin -- regards jl -- regards jl
index problem with write lock
i find one interesting thing. when i index data with 45 solr boxs.(data have 1700w, freebsd6, java: diablo-1.5.0_07-b01, tomcat6), write lock will happen in the procedure. Reindex with solr box which have problem with write block. it show me well. it happen serveral times, so i wanna know why it happen, it should be ok in theory. not every solr box will have this problem. 45 solr boxs have same config and empty index.(i use copy, just rename their directory name) anyone have same problem and know why? -- regards jl
Re: List of highlighted terms from search query
first u try enable highlighting( http://wiki.apache.org/solr/HighlightingParameters) and u try solr admin gui to see its output and u will find what u wanna. 2007/5/23, solruser [EMAIL PROTECTED]: Hi, I am wondering can we get the list of all highlighted terms from the search query. If yes can someone suggest query options? Thanks -- View this message in context: http://www.nabble.com/List-of-highlighted-terms-from-search-query-tf3801344.html#a10755381 Sent from the Solr - User mailing list archive at Nabble.com. -- regards jl
Re: Notice: if u use php and get json style result from solr
the attach is json_encode string which have @ now i find it is PHP JSON function bug because it happen not only when i use the encode string have @. now i use JSON_PHP class.(http://mike.teczno.com/json.html) test code (PHP CODE): ?php require('json.php'); $json = new Services_JSON(); $text = ' {responseHeader:{status:0,QTime:0,params:{fl:* score,start:0,q:postbody:土坡,stylesheet:,solrNum:2,qt:standard,wt:json,rows:10}},response:{numFound:2,start:0,maxScore: 0.2847415,docs:[{id:61410121,posttopic:童年的娱乐 之 漫山遍野飘歌声.精,postbody:漫山遍野飘歌声 我老汉想当年的嗓子,那是相――当的不一般! 清脆?清纯?嘹亮?婉转?甜美?圆润?好像都不能恰如其分地形容我老汉儿时的嗓子(哈哈)! 不信?那我老汉道出一两件往事来,你听了就可以想象得出,我老汉儿时的嗓子是多么的。。多么了。 我老汉的儿时嘛,是上个世纪六十年初到七十年代末的时候了。 就从上小学以后说起吧。 刚上一年级的时候,我们的班主任窦老师就慧眼识英才,发现了我出类拔萃的好嗓子,所以,从一年级到五年级,每年的六一节,我必定登台,亮上一嗓子或者几嗓子――革命现代京剧选段:李玉和《提篮小卖》,少剑波《我们是工农子弟兵》,杨子荣《共产党员》。。 每次登台演唱,都是班主任窦老师拉二胡伴奏。我往台上一站,环顾台下,我们村小学和临村小学一至五年级的学生们早已经屏息而待。窦老师的二胡伴奏过门一响过,我美妙的歌喉即可展开:老乡――!我们是工农子弟兵,来到深山,要消灭反动派,改地换天。。歌喉一亮,清脆,纯正,清越,嘹亮,台下几百双眼睛立刻聚焦于我一身,那眼神,那表情,是惊奇?是羡慕?佩服?是妒忌?一唱完,掌声震耳啊(嘻嘻)! 到了星期六、星期天上山拾草、打柴的时候,就更是我和伙伴们大展歌喉的美好时光了。 在去山坡的路上,我们已经嗓痒难耐。环视近处无人,便放声而歌;到了山上,手割草,嘴放歌,各得其所。 听~罢~奶奶~说~红~~灯~~一声纯正清脆的女声在山坡上陡然响起――是女孩子在唱?非也!我老汉小时候才不稀罕和小丫头片子为伴呢――这是我,以绝对乱真的假嗓唱女声!我的伙伴顺海、方生、硕发,不约而同停了镰刀,拇指一翘:好!而后,齐声接唱:言~语~不多~~道~理~深――!我的独唱就成了交响乐了。 割草累了休息的时候,清唱就变成表演唱了。 那时候,我们刚刚看过了电影故事片《英雄儿女》,影片里的人物、故事���节和插曲,就成了我们表演的保留节目了。 选择一个比四周高起的地方,作为阵地,我,志愿军战士王成冲上阵地,拳头当成了报话器:延安,延安!我是八五一!延安,延安!我是八五一!王成呼叫!王成呼叫! 把镰刀双手平端就成了机枪,左右摆动,嘟嘟。。嘴里摹拟着激烈的射击声,抓起土块、石块当作手榴弹手雷,奋力掷向敌群。。最后,一声高喊:为了胜利,向我开炮――!真正觉得热血沸腾,手握爆破筒,将想象中的引信一拉,一纵身,跳下了土坡!然后,我们伙伴几个站成一排,我,女战士王芳放声歌唱: 风烟滚滚唱英雄,四面青山侧耳听,侧耳听。晴天响雷敲金鼓,大海扬波作和声。人民战士驱虎豹,舍声忘死保和平! 一段唱完,伙伴们齐唱: 为什么战旗美如画?英雄的鲜血染红了她!为什么大地春常在?英雄的生命开鲜花! 山坡上,飘荡着我和伙伴们嘹亮的歌声。 那一回,我们几个伙伴在村南的岭上打猪草。一曲高歌刚罢,忽听一声好!地边的堰墙下一下子冒出来七、八个脑袋!定睛一看,原来是我们村第五生产队的人在下面的地里锄地。这么美妙的歌声竟然被他们白白的偷听了?我一个眼色:转移!伙伴们立马开拔,一会儿我们就把阵地转移到了山坡西面,小小竹排江中游,巍巍青山两岸走。。嘹亮的歌声马上又在山坡上回响。。 上四年级的时候,一天下午的自习课,老师没有来,同学们说的说,闹的闹。我忽然感到嗓痒难耐,于是,在熙熙攘攘的人声里,亮了一嗓子:听~罢~奶奶~说~红~~灯――教室里立刻鸦雀无声,大家纷纷扭头四顾,面面相觑: 哪里放的收音机? 因为歌喉甜美、声名在外、誉满村东了,我们生产小队的几个大姑娘、小媳妇便时常在我上学、放学的路上拦截我,非让我唱一段给她们听不可。 那一天,是在上午放学的路上,我正哼着歌曲往家走,刚拐进冯家胡同,忽然被一个大姑娘和一个小媳妇一前一后堵住了。秋生,唱个歌!要不,别想过去!小媳妇笑嘻嘻地说;快唱吧!大姑娘嘻嘻笑地道。我深悔自己警惕性不高,以致陷如重围!前后瞧瞧,实在走不脱,我只好清清嗓子,嘹亮一声: 风烟滚滚唱英雄,四面青山侧耳听,侧耳听! 我歌喉一展,那大姑娘、小媳妇立刻陶醉在美妙的歌声里;偷眼瞅她们侧耳倾听疏于防范,我拔腿就走,夺路而跑!跑出十几步,听她们在身后喊: 下回逮住你,好好唱给我们听――! 。。 今天就先唱到这里?我们下回再唱?嘿嘿!,siteid:16,sitename:,forumid:5225,forumname:都市夜生活,poster:一点通xylx[274693772],postdt:2006-11-07T19:49:17Z,hasimg:0,postid: [EMAIL PROTECTED],replynum:12,hitnum:179,posturl: http://bbs.qq.com/cgi-bin/bbs/show/content?club=0groupid=104:11005st=sc=messageid=33268 ,score:0.2847415},{id:91446585,posttopic:一个浪子的美国来信(连载) 文/潇水,postbody:写在前面的话 因为我在国外偷闲写了一点不好看的东西,就在这样的夏暮季节,坐立不安起来,希望多被别人来读它们。然而这个想法却是很要不得的,因为我的从前大学教育,都集中在了麦克斯韦方程和电子技术方面,纵观我的历史,一直没能沾染什么文化和文字。 我最早出生,是在河北一个有山坡和树林河水的小地方,我爸妈虽然都是好教书人,但其实更关心是把家庭生计搞得不艰苦一点。所以他们那年代我生在72年就于院子里垒了鸡窝以饲养鸡和兔子。我少年闲余,就给小兔们弄青草吃,做这一类有趣的事业,并不徜徉于优秀书籍。直到进了清华念书以后,才开始窥测到一些人间所公认必读的好书,比如什么什么的。但更主要的时间是在做高压放电实验之类的事业。 后来我就呆在了北京,在外国公司里做事。和我经常来往的人,简约平朴,相与友善的,是一个甘肃宕昌籍的叫杨向荣的人,大学时期,常散步聊天,于水木清华的池木土坡一带,共同望一望夜空里的星光幽弱颤抖。 不久,我到国外读书,离纽约很近的一个大学城里,僻静无为,就偶于月明气清、夜色淡荡时刻,作一封信去,把这边的事写给杨向荣知道。积累久了,就得了好几封。 最近因为无聊,就从过年以后,一直在整理这些旧信。信整理得很不好,到了今天的深夜,也没有搞得多少的迹像。我想这些旧信也许在英流网上可以有用,就把它的完成的部分,弄在下面,以显示自己格外地敝帚自珍。 --2000-09-04|--|1.乘钻天猴出埃及记 向荣兄: 今天翻旧书本,发现扉页上有几行写于去年秋天北京的感想,权且抄在这里-- 昏黄漠漠的阳台往外,人生景致的远处,祖国的秋天正在节节垂下,从古长城到丝绸路,从火焰山到太行山,我的无数坏情绪,正随着暖气和水,随着下面的喧嚣叫卖,在城市的金属管子里被纷纷敲响。那些愚蠢的青年时代、神经衰弱的大学生涯,正在秋色一张大席的掩护下,纷纷登上忽明忽暗的餐桌与餐器。 在各自的角落里发光发热,在各自的角落里发愁,在各自的角落里把秋天的大网挖坏一角。 我与其说与秋天的蟋蟀角斗虫唱,不如说仍是在挖土,在挖土的过程中神色狡黠、心情不定,像一匹惯于白昼飞驰的马,在预感的夜幕边缘,沿着荒原静静地惊走。 --南苑机场北门以北五百米,赁居的楼上,十月 这段文字,全是空洞的话。 又过了一两个月,在秋天的大网上空费了许多功夫,而转过年初,新泽西州立大学的通知书就来了,时间紧迫,要马上动身。我在这荒远的的南郊机场北门,住得也太久了,该是离开父母之邦的时刻了。我弟弟和母亲拖着两包行李往北边的首都机场去送我,我父亲病卧不能出行。 这样,在天色阴暗的黎明,我推着行李从机场的小口进去,母弟二��依在小口的冬天的铁栏上,痴痴地向渐渐混入人群的我望,渐渐地望不见我了,连我的行李也望不见了。 而我的眼泪,就几乎要淌下来了。 我上了飞机,人们互不说话,只听见空调咝咝鼓气的声音。 想说出什么,想笑出一笑,却没有。 不久,飞机突然失去理智地猛跑,不由分说跳上了天,我只好随着它疯了地飞,暂搁下我的心情不能管。 服务人员渐渐开始讲话,话声也渐渐入了我的大脑,把我从那离愁情绪中往外拔。 飞了一小时,好似一昼夜一样烦乱漫长,然后突然落在上海,人们纷纷下飞机,我不明就里,也裹着下了飞机,以为此次旅行就这样结束了--兴许在我刚才飞行的时候,国家出了事,改了主意,冻结一切出国人员了。我也不打听,心想不出国也好,依旧回北京南郊的楼上去,并且几乎开始谋划回去之后的生活了。 然后有官员领着我们坐等,我不急,心想政府总会管好我们的。果然,开始排队交验护照,轮到我时,我想,她不会突然说我的护照是假的吧,或者她突然说,护照是真的,但我这人是假的。但她终于什么也没有说,给计算机输入了点什么,往护照上加了个戳,就还了我,示意我往前走,我还觉得兴许有点什么事不妥,但不待想,人就已这样出了我国的海关了。 夹在旅客之中,又上了原先的飞机,上海的早晨静悄悄的,两年前我曾经游过这里,带着对江左的好奇和少年情绪。胡思乱想之际,飞机忽又自作主张地猛跑,机身一摇,双脚跳上空空的天。 啊,好大好大的蓝天,好蓝好蓝的大天。 空中服务小姐们是纯净而勤快的,好似一把剪刀裁出,她们扶着小车一个一个为大家斟饮料。而大家各持自己的面孔、自己的愁事和自己的钱财、自己的来路与去程,不露声色,只喝饮料,似乎饮料正在高空中解决着人们的精神需求。 我的前方挂着一台小电视,放一些令人费神的画片,有时出现一张地图,从图上看出我们的飞机正在广阔的太平洋上吃力地吞食着路程,刚刚移到日本国的上空。 我往下望了望,似乎也没有感到地气的变化。半天云团像电影院里的一片观众,仰着头看银幕上我们的飞机飞。 一瞬间,我似乎也在电影上了,导演给我们每个乘客都安排了故事与命运。坐在我一旁的这个人,他的脑门上刻有皱纹,松驰的脸颊对一切漠不经心,他该是个背运的家伙?还是时代的宠儿?他…… 我戴上耳机,,siteid:3583,sitename:,forumid:42603,forumname:网文原创,poster:美丽土豆001,postdt:2007-03-10T20:24:00Z,hasimg:0,postid: [EMAIL PROTECTED],replynum:2,hitnum:19,posturl: http://bbs.runsky.com/bbs/thread-2807543-1-1.html,score:0.2847415}]}} '; var_dump(json_decode($text)); var_dump($json-decode($text)); ? 2007/5/19, Chris Hostetter [EMAIL PROTECTED]: : I don't see
Notice: if u use php and get json style result from solr
if u get null from json_decode($data),,,maybe ur $data have '@'fix way is replace it before u do json_decode i try json_encode with php and json_decode with phpit is no problem when i use '@' maybe it only happen encode(by java) and decode(by php) -- regards jl
Re: Notice: if u use php and get json style result from solr
2007/5/18, Chris Hostetter [EMAIL PROTECTED]: : : if u get null from json_decode($data),,,maybe ur $data have '@'fix way : is replace it before u do json_decode : : i try json_encode with php and json_decode with phpit is no problem when : i use '@' : : maybe it only happen encode(by java) and decode(by php) this may be a bug in our JSONResonseWRiter ... is @ a special character in JSON that needs to be escaped? I don't see anything about it in RFC 4627 (which seems to be the most current JSON spec) so perhaps this is just a bug in the PHP json parser? i try php with encode and decode. It works well although it has @ so i don't think it is php's bug -Hoss -- regards jl
Re: PriceJunkie.com using solr!
how many solr instance? 2007/5/17, Yonik Seeley [EMAIL PROTECTED]: Congrats, very nice job! It's fast too. -Yonik On 5/16/07, Mike Austin [EMAIL PROTECTED] wrote: I just wanted to say thanks to everyone for the creation of solr. I've been using it for a while now and I have recently brought one of my side projects online. I have several other projects that will be using solr for it's search and facets. Please check out www.pricejunkie.com and let us know what you think.. You can give feedback and/or sign up on the mailing list for future updates. The site is very basic right now and many new and useful features plus merchants and product categories will be coming soon! I thought it would be a good idea to at least have a few people use it to get some feedback early and often. Some of the nice things behind the scenes that we did with solr: - created custom request handlers that have category to facet to attribute caching built in - category to facet management - ability to manage facet groups (attributes within a set facet) and assign them to categories - ability to create any category structure and share facet groups - facet inheritance for any category (a facet group can be defined on a parent category and pushed down to all children) - ability to create sub-categories as facets instead of normal sub categories - simple xml configuration for the final outputted category configuration file I'm sure there are more cool things but that is all for now. Join the mailing list to see more improvements in the future. Also.. how do I get added to the Using Solr wiki page? Thanks, Mike Austin -- regards jl
Question: Pagination with multi index box
if use multi index box, how to pagination with sort by score correctly? for example, i wanna query search with 60 index box and sort by score. i don't know the num found from every index box which have different content. if promise 10 page with sort score correctly, i think solr 's start is 0, and rows is 100.(10 result per page) 60*100=6000, sort it and get top 100 to cache. it is very slove although it promise 10 page with sort score correctly. any idea to fix it? fast and correct. -- regards jl
Re: Question: Pagination with multi index box
2007/5/15, Mike Klaas [EMAIL PROTECTED]: On 14-May-07, at 1:35 AM, James liu wrote: if use multi index box, how to pagination with sort by score correctly? for example, i wanna query search with 60 index box and sort by score. i don't know the num found from every index box which have different content. if promise 10 page with sort score correctly, i think solr 's start is 0, and rows is 100.(10 result per page) 60*100=6000, sort it and get top 100 to cache. it is very slove although it promise 10 page with sort score correctly. With few index partitions, you it is sufficient to ask for startAt +numNeeded docs from each partition and sort globally. Normally if you wanted 10 for the first page, you would ask for 10 from each server and cache the remainder. It is better to ask for more later if the user asks for page ten. When you get up to 60 partitions, you should make it a multi stage process. Assuming your partitions are disjoint and evenly distributed, estimate the number of documents that will appear in the final result from each. yes, partitions distrbuted. Double or triple that (and put a minimum threshold), try to assemble the number of documents you require, and if one partition runs out of docs before it is done, request a new round. i dont' know what u mean runs out one user request will generate 60 partitions request. they work in parallel。 so i don't know every partion's status before they done. To promise 10 page result sorted by score correctly, the only way seems to get 100 results(rows=100) from each partitioin. but it very slow. now i wanna find a way to get result sorted by score correctly and search fast. -Mike Thks Mike. But it not i want. -- regards jl
Re: Question: Pagination with multi index box
if i set rows=(page-1)*10,,,it will lose more result which fits query. how to set start when pagination. 2007/5/15, James liu [EMAIL PROTECTED]: 2007/5/15, Mike Klaas [EMAIL PROTECTED]: On 14-May-07, at 1:35 AM, James liu wrote: if use multi index box, how to pagination with sort by score correctly? for example, i wanna query search with 60 index box and sort by score. i don't know the num found from every index box which have different content. if promise 10 page with sort score correctly, i think solr 's start is 0, and rows is 100.(10 result per page) 60*100=6000, sort it and get top 100 to cache. it is very slove although it promise 10 page with sort score correctly. With few index partitions, you it is sufficient to ask for startAt +numNeeded docs from each partition and sort globally. Normally if you wanted 10 for the first page, you would ask for 10 from each server and cache the remainder. It is better to ask for more later if the user asks for page ten. When you get up to 60 partitions, you should make it a multi stage process. Assuming your partitions are disjoint and evenly distributed, estimate the number of documents that will appear in the final result from each. yes, partitions distrbuted. Double or triple that (and put a minimum threshold), try to assemble the number of documents you require, and if one partition runs out of docs before it is done, request a new round. i dont' know what u mean runs out one user request will generate 60 partitions request. they work in parallel。 so i don't know every partion's status before they done. To promise 10 page result sorted by score correctly, the only way seems to get 100 results(rows=100) from each partitioin. but it very slow. now i wanna find a way to get result sorted by score correctly and search fast. -Mike Thks Mike. But it not i want. -- regards jl -- regards jl