No problem!  It's exciting :)

Mike

On Tue, Oct 6, 2009 at 10:40 AM, Uwe Schindler <u...@thetaphi.de> wrote:
> Thanks, sorry for extra work! I missed to do this after the svn copy :(
>
> -----
> Uwe Schindler
> H.-H.-Meier-Allee 63, D-28213 Bremen
> http://www.thetaphi.de
> eMail: u...@thetaphi.de
>
>
>> -----Original Message-----
>> From: Michael McCandless [mailto:luc...@mikemccandless.com]
>> Sent: Tuesday, October 06, 2009 4:37 PM
>> To: java-dev@lucene.apache.org
>> Subject: Re: svn commit: r822284 -
>> /lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibi
>> lity.java
>>
>> OK will do.
>>
>> Mike
>>
>> On Tue, Oct 6, 2009 at 10:23 AM, Uwe Schindler <u...@thetaphi.de> wrote:
>> > Can you add this patch to backwards, too? I forgot, that some of the
>> > backwards-changes also applied to BW, but for completeness, not sure, if
>> a
>> > tag is also needed.
>> >
>> > -----
>> > Uwe Schindler
>> > H.-H.-Meier-Allee 63, D-28213 Bremen
>> > http://www.thetaphi.de
>> > eMail: u...@thetaphi.de
>> >
>> >
>> >> -----Original Message-----
>> >> From: mikemcc...@apache.org [mailto:mikemcc...@apache.org]
>> >> Sent: Tuesday, October 06, 2009 4:13 PM
>> >> To: java-comm...@lucene.apache.org
>> >> Subject: svn commit: r822284 -
>> >>
>> /lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibi
>> >> lity.java
>> >>
>> >> Author: mikemccand
>> >> Date: Tue Oct  6 14:12:46 2009
>> >> New Revision: 822284
>> >>
>> >> URL: http://svn.apache.org/viewvc?rev=822284&view=rev
>> >> Log:
>> >> fix TestBackwardsCompability to not use deprecated APIs
>> >>
>> >> Modified:
>> >>
>> >>
>> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil
>> >> ity.java
>> >>
>> >> Modified:
>> >>
>> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil
>> >> ity.java
>> >> URL:
>> >>
>> http://svn.apache.org/viewvc/lucene/java/trunk/src/test/org/apache/lucene/
>> >>
>> index/TestBackwardsCompatibility.java?rev=822284&r1=822283&r2=822284&view=
>> >> diff
>> >>
>> ==========================================================================
>> >> ====
>> >> ---
>> >>
>> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil
>> >> ity.java (original)
>> >> +++
>> >>
>> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil
>> >> ity.java Tue Oct  6 14:12:46 2009
>> >> @@ -158,11 +158,7 @@
>> >>      for(int i=0;i<oldNames.length;i++) {
>> >>        String dirName = "src/test/org/apache/lucene/index/index." +
>> >> oldNames[i];
>> >>        unzip(dirName, oldNames[i]);
>> >> -      changeIndexNoAdds(oldNames[i], true);
>> >> -      rmDir(oldNames[i]);
>> >> -
>> >> -      unzip(dirName, oldNames[i]);
>> >> -      changeIndexNoAdds(oldNames[i], false);
>> >> +      changeIndexNoAdds(oldNames[i]);
>> >>        rmDir(oldNames[i]);
>> >>      }
>> >>    }
>> >> @@ -171,11 +167,7 @@
>> >>      for(int i=0;i<oldNames.length;i++) {
>> >>        String dirName = "src/test/org/apache/lucene/index/index." +
>> >> oldNames[i];
>> >>        unzip(dirName, oldNames[i]);
>> >> -      changeIndexWithAdds(oldNames[i], true);
>> >> -      rmDir(oldNames[i]);
>> >> -
>> >> -      unzip(dirName, oldNames[i]);
>> >> -      changeIndexWithAdds(oldNames[i], false);
>> >> +      changeIndexWithAdds(oldNames[i]);
>> >>        rmDir(oldNames[i]);
>> >>      }
>> >>    }
>> >> @@ -196,7 +188,7 @@
>> >>      dirName = fullDir(dirName);
>> >>
>> >>      Directory dir = FSDirectory.open(new File(dirName));
>> >> -    IndexSearcher searcher = new IndexSearcher(dir);
>> >> +    IndexSearcher searcher = new IndexSearcher(dir, true);
>> >>      IndexReader reader = searcher.getIndexReader();
>> >>
>> >>      _TestUtil.checkIndex(dir);
>> >> @@ -267,14 +259,14 @@
>> >>
>> >>    /* Open pre-lockless index, add docs, do a delete &
>> >>     * setNorm, and search */
>> >> -  public void changeIndexWithAdds(String dirName, boolean autoCommit)
>> >> throws IOException {
>> >> +  public void changeIndexWithAdds(String dirName) throws IOException {
>> >>      String origDirName = dirName;
>> >>      dirName = fullDir(dirName);
>> >>
>> >>      Directory dir = FSDirectory.open(new File(dirName));
>> >>
>> >>      // open writer
>> >> -    IndexWriter writer = new IndexWriter(dir, autoCommit, new
>> >> WhitespaceAnalyzer(), false);
>> >> +    IndexWriter writer = new IndexWriter(dir, new
>> WhitespaceAnalyzer(),
>> >> false, IndexWriter.MaxFieldLength.UNLIMITED);
>> >>
>> >>      // add 10 docs
>> >>      for(int i=0;i<10;i++) {
>> >> @@ -288,11 +280,11 @@
>> >>      } else {
>> >>        expected = 46;
>> >>      }
>> >> -    assertEquals("wrong doc count", expected, writer.docCount());
>> >> +    assertEquals("wrong doc count", expected, writer.maxDoc());
>> >>      writer.close();
>> >>
>> >>      // make sure searching sees right # hits
>> >> -    IndexSearcher searcher = new IndexSearcher(dir);
>> >> +    IndexSearcher searcher = new IndexSearcher(dir, true);
>> >>      ScoreDoc[] hits = searcher.search(new TermQuery(new
>> Term("content",
>> >> "aaa")), null, 1000).scoreDocs;
>> >>      Document d = searcher.doc(hits[0].doc);
>> >>      assertEquals("wrong first document", "21", d.get("id"));
>> >> @@ -301,7 +293,7 @@
>> >>
>> >>      // make sure we can do delete & setNorm against this
>> >>      // pre-lockless segment:
>> >> -    IndexReader reader = IndexReader.open(dir);
>> >> +    IndexReader reader = IndexReader.open(dir, false);
>> >>      Term searchTerm = new Term("id", "6");
>> >>      int delCount = reader.deleteDocuments(searchTerm);
>> >>      assertEquals("wrong delete count", 1, delCount);
>> >> @@ -309,7 +301,7 @@
>> >>      reader.close();
>> >>
>> >>      // make sure they "took":
>> >> -    searcher = new IndexSearcher(dir);
>> >> +    searcher = new IndexSearcher(dir, true);
>> >>      hits = searcher.search(new TermQuery(new Term("content", "aaa")),
>> >> null, 1000).scoreDocs;
>> >>      assertEquals("wrong number of hits", 43, hits.length);
>> >>      d = searcher.doc(hits[0].doc);
>> >> @@ -318,11 +310,11 @@
>> >>      searcher.close();
>> >>
>> >>      // optimize
>> >> -    writer = new IndexWriter(dir, autoCommit, new
>> WhitespaceAnalyzer(),
>> >> false);
>> >> +    writer = new IndexWriter(dir, new WhitespaceAnalyzer(), false,
>> >> IndexWriter.MaxFieldLength.UNLIMITED);
>> >>      writer.optimize();
>> >>      writer.close();
>> >>
>> >> -    searcher = new IndexSearcher(dir);
>> >> +    searcher = new IndexSearcher(dir, true);
>> >>      hits = searcher.search(new TermQuery(new Term("content", "aaa")),
>> >> null, 1000).scoreDocs;
>> >>      assertEquals("wrong number of hits", 43, hits.length);
>> >>      d = searcher.doc(hits[0].doc);
>> >> @@ -335,14 +327,14 @@
>> >>
>> >>    /* Open pre-lockless index, add docs, do a delete &
>> >>     * setNorm, and search */
>> >> -  public void changeIndexNoAdds(String dirName, boolean autoCommit)
>> >> throws IOException {
>> >> +  public void changeIndexNoAdds(String dirName) throws IOException {
>> >>
>> >>      dirName = fullDir(dirName);
>> >>
>> >>      Directory dir = FSDirectory.open(new File(dirName));
>> >>
>> >>      // make sure searching sees right # hits
>> >> -    IndexSearcher searcher = new IndexSearcher(dir);
>> >> +    IndexSearcher searcher = new IndexSearcher(dir, true);
>> >>      ScoreDoc[] hits = searcher.search(new TermQuery(new
>> Term("content",
>> >> "aaa")), null, 1000).scoreDocs;
>> >>      assertEquals("wrong number of hits", 34, hits.length);
>> >>      Document d = searcher.doc(hits[0].doc);
>> >> @@ -351,7 +343,7 @@
>> >>
>> >>      // make sure we can do a delete & setNorm against this
>> >>      // pre-lockless segment:
>> >> -    IndexReader reader = IndexReader.open(dir);
>> >> +    IndexReader reader = IndexReader.open(dir, false);
>> >>      Term searchTerm = new Term("id", "6");
>> >>      int delCount = reader.deleteDocuments(searchTerm);
>> >>      assertEquals("wrong delete count", 1, delCount);
>> >> @@ -359,7 +351,7 @@
>> >>      reader.close();
>> >>
>> >>      // make sure they "took":
>> >> -    searcher = new IndexSearcher(dir);
>> >> +    searcher = new IndexSearcher(dir, true);
>> >>      hits = searcher.search(new TermQuery(new Term("content", "aaa")),
>> >> null, 1000).scoreDocs;
>> >>      assertEquals("wrong number of hits", 33, hits.length);
>> >>      d = searcher.doc(hits[0].doc);
>> >> @@ -368,11 +360,11 @@
>> >>      searcher.close();
>> >>
>> >>      // optimize
>> >> -    IndexWriter writer = new IndexWriter(dir, autoCommit, new
>> >> WhitespaceAnalyzer(), false);
>> >> +    IndexWriter writer = new IndexWriter(dir, new
>> WhitespaceAnalyzer(),
>> >> false, IndexWriter.MaxFieldLength.UNLIMITED);
>> >>      writer.optimize();
>> >>      writer.close();
>> >>
>> >> -    searcher = new IndexSearcher(dir);
>> >> +    searcher = new IndexSearcher(dir, true);
>> >>      hits = searcher.search(new TermQuery(new Term("content", "aaa")),
>> >> null, 1000).scoreDocs;
>> >>      assertEquals("wrong number of hits", 33, hits.length);
>> >>      d = searcher.doc(hits[0].doc);
>> >> @@ -397,7 +389,7 @@
>> >>      for(int i=0;i<35;i++) {
>> >>        addDoc(writer, i);
>> >>      }
>> >> -    assertEquals("wrong doc count", 35, writer.docCount());
>> >> +    assertEquals("wrong doc count", 35, writer.maxDoc());
>> >>      writer.close();
>> >>
>> >>      // open fresh writer so we get no prx file in the added segment
>> >> @@ -408,7 +400,7 @@
>> >>      writer.close();
>> >>
>> >>      // Delete one doc so we get a .del file:
>> >> -    IndexReader reader = IndexReader.open(dir);
>> >> +    IndexReader reader = IndexReader.open(dir, false);
>> >>      Term searchTerm = new Term("id", "7");
>> >>      int delCount = reader.deleteDocuments(searchTerm);
>> >>      assertEquals("didn't delete the right number of documents", 1,
>> >> delCount);
>> >> @@ -422,71 +414,66 @@
>> >>
>> >>    public void testExactFileNames() throws IOException {
>> >>
>> >> -    for(int pass=0;pass<2;pass++) {
>> >> -
>> >> -      String outputDir = "lucene.backwardscompat0.index";
>> >> -      rmDir(outputDir);
>> >> +    String outputDir = "lucene.backwardscompat0.index";
>> >> +    rmDir(outputDir);
>> >>
>> >> -      try {
>> >> -        Directory dir = FSDirectory.open(new
>> File(fullDir(outputDir)));
>> >> +    try {
>> >> +      Directory dir = FSDirectory.open(new File(fullDir(outputDir)));
>> >>
>> >> -        boolean autoCommit = 0 == pass;
>> >> -
>> >> -        IndexWriter writer = new IndexWriter(dir, autoCommit, new
>> >> WhitespaceAnalyzer(), true);
>> >> -        writer.setRAMBufferSizeMB(16.0);
>> >> -        for(int i=0;i<35;i++) {
>> >> -          addDoc(writer, i);
>> >> -        }
>> >> -        assertEquals("wrong doc count", 35, writer.docCount());
>> >> -        writer.close();
>> >> +      IndexWriter writer = new IndexWriter(dir, new
>> WhitespaceAnalyzer(),
>> >> true, IndexWriter.MaxFieldLength.UNLIMITED);
>> >> +      writer.setRAMBufferSizeMB(16.0);
>> >> +      for(int i=0;i<35;i++) {
>> >> +        addDoc(writer, i);
>> >> +      }
>> >> +      assertEquals("wrong doc count", 35, writer.maxDoc());
>> >> +      writer.close();
>> >>
>> >> -        // Delete one doc so we get a .del file:
>> >> -        IndexReader reader = IndexReader.open(dir);
>> >> -        Term searchTerm = new Term("id", "7");
>> >> -        int delCount = reader.deleteDocuments(searchTerm);
>> >> -        assertEquals("didn't delete the right number of documents", 1,
>> >> delCount);
>> >> -
>> >> -        // Set one norm so we get a .s0 file:
>> >> -        reader.setNorm(21, "content", (float) 1.5);
>> >> -        reader.close();
>> >> -
>> >> -        // The numbering of fields can vary depending on which
>> >> -        // JRE is in use.  On some JREs we see content bound to
>> >> -        // field 0; on others, field 1.  So, here we have to
>> >> -        // figure out which field number corresponds to
>> >> -        // "content", and then set our expected file names below
>> >> -        // accordingly:
>> >> -        CompoundFileReader cfsReader = new CompoundFileReader(dir,
>> >> "_0.cfs");
>> >> -        FieldInfos fieldInfos = new FieldInfos(cfsReader, "_0.fnm");
>> >> -        int contentFieldIndex = -1;
>> >> -        for(int i=0;i<fieldInfos.size();i++) {
>> >> -          FieldInfo fi = fieldInfos.fieldInfo(i);
>> >> -          if (fi.name.equals("content")) {
>> >> -            contentFieldIndex = i;
>> >> -            break;
>> >> -          }
>> >> +      // Delete one doc so we get a .del file:
>> >> +      IndexReader reader = IndexReader.open(dir, false);
>> >> +      Term searchTerm = new Term("id", "7");
>> >> +      int delCount = reader.deleteDocuments(searchTerm);
>> >> +      assertEquals("didn't delete the right number of documents", 1,
>> >> delCount);
>> >> +
>> >> +      // Set one norm so we get a .s0 file:
>> >> +      reader.setNorm(21, "content", (float) 1.5);
>> >> +      reader.close();
>> >> +
>> >> +      // The numbering of fields can vary depending on which
>> >> +      // JRE is in use.  On some JREs we see content bound to
>> >> +      // field 0; on others, field 1.  So, here we have to
>> >> +      // figure out which field number corresponds to
>> >> +      // "content", and then set our expected file names below
>> >> +      // accordingly:
>> >> +      CompoundFileReader cfsReader = new CompoundFileReader(dir,
>> >> "_0.cfs");
>> >> +      FieldInfos fieldInfos = new FieldInfos(cfsReader, "_0.fnm");
>> >> +      int contentFieldIndex = -1;
>> >> +      for(int i=0;i<fieldInfos.size();i++) {
>> >> +        FieldInfo fi = fieldInfos.fieldInfo(i);
>> >> +        if (fi.name.equals("content")) {
>> >> +          contentFieldIndex = i;
>> >> +          break;
>> >>          }
>> >> -        cfsReader.close();
>> >> -        assertTrue("could not locate the 'content' field number in the
>> >> _2.cfs segment", contentFieldIndex != -1);
>> >> +      }
>> >> +      cfsReader.close();
>> >> +      assertTrue("could not locate the 'content' field number in the
>> >> _2.cfs segment", contentFieldIndex != -1);
>> >>
>> >> -        // Now verify file names:
>> >> -        String[] expected;
>> >> -        expected = new String[] {"_0.cfs",
>> >> -                    "_0_1.del",
>> >> -                    "_0_1.s" + contentFieldIndex,
>> >> -                    "segments_3",
>> >> -                    "segments.gen"};
>> >> -
>> >> -        String[] actual = dir.listAll();
>> >> -        Arrays.sort(expected);
>> >> -        Arrays.sort(actual);
>> >> -        if (!Arrays.equals(expected, actual)) {
>> >> -          fail("incorrect filenames in index: expected:\n    " +
>> >> asString(expected) + "\n  actual:\n    " + asString(actual));
>> >> -        }
>> >> -        dir.close();
>> >> -      } finally {
>> >> -        rmDir(outputDir);
>> >> +      // Now verify file names:
>> >> +      String[] expected;
>> >> +      expected = new String[] {"_0.cfs",
>> >> +                               "_0_1.del",
>> >> +                               "_0_1.s" + contentFieldIndex,
>> >> +                               "segments_3",
>> >> +                               "segments.gen"};
>> >> +
>> >> +      String[] actual = dir.listAll();
>> >> +      Arrays.sort(expected);
>> >> +      Arrays.sort(actual);
>> >> +      if (!Arrays.equals(expected, actual)) {
>> >> +        fail("incorrect filenames in index: expected:\n    " +
>> >> asString(expected) + "\n  actual:\n    " + asString(actual));
>> >>        }
>> >> +      dir.close();
>> >> +    } finally {
>> >> +      rmDir(outputDir);
>> >>      }
>> >>    }
>> >>
>> >> @@ -509,17 +496,17 @@
>> >>      doc.add(new Field("autf8", "Lu\uD834\uDD1Ece\uD834\uDD60ne \u0000
>> >> \u2620 ab\ud917\udc17cd", Field.Store.YES, Field.Index.ANALYZED,
>> >> Field.TermVector.WITH_POSITIONS_OFFSETS));
>> >>      doc.add(new Field("utf8", "Lu\uD834\uDD1Ece\uD834\uDD60ne \u0000
>> >> \u2620 ab\ud917\udc17cd", Field.Store.YES, Field.Index.ANALYZED,
>> >> Field.TermVector.WITH_POSITIONS_OFFSETS));
>> >>      doc.add(new Field("content2", "here is more content with aaa aaa
>> >> aaa", Field.Store.YES, Field.Index.ANALYZED,
>> >> Field.TermVector.WITH_POSITIONS_OFFSETS));
>> >> -    doc.add(new Field("fie\u2C77ld", "field with non-ascii name",
>> >> Field.Store.YES, Field.Index.TOKENIZED,
>> >> Field.TermVector.WITH_POSITIONS_OFFSETS));
>> >> +    doc.add(new Field("fie\u2C77ld", "field with non-ascii name",
>> >> Field.Store.YES, Field.Index.ANALYZED,
>> >> Field.TermVector.WITH_POSITIONS_OFFSETS));
>> >>      writer.addDocument(doc);
>> >>    }
>> >>
>> >>    private void addNoProxDoc(IndexWriter writer) throws IOException {
>> >>      Document doc = new Document();
>> >>      Field f = new Field("content3", "aaa", Field.Store.YES,
>> >> Field.Index.ANALYZED);
>> >> -    f.setOmitTf(true);
>> >> +    f.setOmitTermFreqAndPositions(true);
>> >>      doc.add(f);
>> >>      f = new Field("content4", "aaa", Field.Store.YES, Field.Index.NO);
>> >> -    f.setOmitTf(true);
>> >> +    f.setOmitTermFreqAndPositions(true);
>> >>      doc.add(f);
>> >>      writer.addDocument(doc);
>> >>    }
>> >>
>> >
>> >
>> >
>> > ---------------------------------------------------------------------
>> > To unsubscribe, e-mail: java-dev-unsubscr...@lucene.apache.org
>> > For additional commands, e-mail: java-dev-h...@lucene.apache.org
>> >
>> >
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: java-dev-unsubscr...@lucene.apache.org
>> For additional commands, e-mail: java-dev-h...@lucene.apache.org
>
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: java-dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: java-dev-h...@lucene.apache.org
>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: java-dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: java-dev-h...@lucene.apache.org

Reply via email to