/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ using System; using NUnit.Framework; using Analyzer = Lucene.Net.Analysis.Analyzer; using StandardAnalyzer = Lucene.Net.Analysis.Standard.StandardAnalyzer; using Document = Lucene.Net.Documents.Document; using Field = Lucene.Net.Documents.Field; using Index = Lucene.Net.Documents.Field.Index; using Store = Lucene.Net.Documents.Field.Store; using Norm = Lucene.Net.Index.SegmentReader.Norm; using Directory = Lucene.Net.Store.Directory; using FSDirectory = Lucene.Net.Store.FSDirectory; using MockRAMDirectory = Lucene.Net.Store.MockRAMDirectory; using DefaultSimilarity = Lucene.Net.Search.DefaultSimilarity; using Similarity = Lucene.Net.Search.Similarity; using LuceneTestCase = Lucene.Net.Util.LuceneTestCase; namespace Lucene.Net.Index { /// Tests cloning IndexReader norms [TestFixture] public class TestIndexReaderCloneNorms:LuceneTestCase { [Serializable] private class SimilarityOne:DefaultSimilarity { public SimilarityOne(TestIndexReaderCloneNorms enclosingInstance) { InitBlock(enclosingInstance); } private void InitBlock(TestIndexReaderCloneNorms enclosingInstance) { this.enclosingInstance = enclosingInstance; } private TestIndexReaderCloneNorms enclosingInstance; public TestIndexReaderCloneNorms Enclosing_Instance { get { return enclosingInstance; } } public override float LengthNorm(System.String fieldName, int numTerms) { return 1; } } private const int NUM_FIELDS = 10; private Similarity similarityOne; private Analyzer anlzr; private int numDocNorms; private System.Collections.ArrayList norms; private System.Collections.ArrayList modifiedNorms; private float lastNorm = 0; private float normDelta = (float) 0.001; public TestIndexReaderCloneNorms(System.String s):base(s) { } public TestIndexReaderCloneNorms(): base("") { } [SetUp] public override void SetUp() { base.SetUp(); similarityOne = new SimilarityOne(this); anlzr = new StandardAnalyzer(); } /// Test that norms values are preserved as the index is maintained. Including /// separate norms. Including merging indexes with seprate norms. Including /// optimize. /// [Test] public virtual void TestNorms() { // tmp dir System.String tempDir = System.IO.Path.GetTempPath(); if (tempDir == null) { throw new System.IO.IOException("java.io.tmpdir undefined, cannot run test"); } // test with a single index: index1 System.IO.FileInfo indexDir1 = new System.IO.FileInfo(System.IO.Path.Combine(tempDir, "lucenetestindex1")); Directory dir1 = FSDirectory.Open(indexDir1); IndexWriter.Unlock(dir1); norms = new System.Collections.ArrayList(); modifiedNorms = new System.Collections.ArrayList(); CreateIndex(dir1); DoTestNorms(dir1); // test with a single index: index2 System.Collections.ArrayList norms1 = norms; System.Collections.ArrayList modifiedNorms1 = modifiedNorms; int numDocNorms1 = numDocNorms; norms = new System.Collections.ArrayList(); modifiedNorms = new System.Collections.ArrayList(); numDocNorms = 0; System.IO.FileInfo indexDir2 = new System.IO.FileInfo(System.IO.Path.Combine(tempDir, "lucenetestindex2")); Directory dir2 = FSDirectory.Open(indexDir2); CreateIndex(dir2); DoTestNorms(dir2); // add index1 and index2 to a third index: index3 System.IO.FileInfo indexDir3 = new System.IO.FileInfo(System.IO.Path.Combine(tempDir, "lucenetestindex3")); Directory dir3 = FSDirectory.Open(indexDir3); CreateIndex(dir3); IndexWriter iw = new IndexWriter(dir3, anlzr, false, IndexWriter.MaxFieldLength.LIMITED); iw.SetMaxBufferedDocs(5); iw.SetMergeFactor(3); iw.AddIndexes(new Directory[]{dir1, dir2}); iw.Close(); norms1.AddRange(norms); norms = norms1; modifiedNorms1.AddRange(modifiedNorms); modifiedNorms = modifiedNorms1; numDocNorms += numDocNorms1; // test with index3 VerifyIndex(dir3); DoTestNorms(dir3); // now with optimize iw = new IndexWriter(dir3, anlzr, false, IndexWriter.MaxFieldLength.LIMITED); iw.SetMaxBufferedDocs(5); iw.SetMergeFactor(3); iw.Optimize(); iw.Close(); VerifyIndex(dir3); dir1.Close(); dir2.Close(); dir3.Close(); } // try cloning and reopening the norms private void DoTestNorms(Directory dir) { AddDocs(dir, 12, true); IndexReader ir = IndexReader.Open(dir); VerifyIndex(ir); ModifyNormsForF1(ir); IndexReader irc = (IndexReader) ir.Clone(); // IndexReader.open(dir);//ir.clone(); VerifyIndex(irc); ModifyNormsForF1(irc); IndexReader irc3 = (IndexReader) irc.Clone(); VerifyIndex(irc3); ModifyNormsForF1(irc3); VerifyIndex(irc3); irc3.Flush(); irc3.Close(); } [Test] public virtual void TestNormsClose() { Directory dir1 = new MockRAMDirectory(); TestIndexReaderReopen.CreateIndex(dir1, false); SegmentReader reader1 = SegmentReader.GetOnlySegmentReader(dir1); reader1.Norms("field1"); Norm r1norm = (Norm) reader1.norms_ForNUnit["field1"]; SegmentReader.Ref r1BytesRef = r1norm.BytesRef(); SegmentReader reader2 = (SegmentReader) reader1.Clone(); Assert.AreEqual(2, r1norm.BytesRef().RefCount()); reader1.Close(); Assert.AreEqual(1, r1BytesRef.RefCount()); reader2.Norms("field1"); reader2.Close(); dir1.Close(); } [Test] public virtual void TestNormsRefCounting() { Directory dir1 = new MockRAMDirectory(); TestIndexReaderReopen.CreateIndex(dir1, false); IndexReader reader1 = IndexReader.Open(dir1); IndexReader reader2C = (IndexReader) reader1.Clone(); SegmentReader segmentReader2C = SegmentReader.GetOnlySegmentReader(reader2C); segmentReader2C.Norms("field1"); // load the norms for the field Norm reader2CNorm = (Norm) segmentReader2C.norms_ForNUnit["field1"]; Assert.IsTrue(reader2CNorm.BytesRef().RefCount() == 2, "reader2CNorm.bytesRef()=" + reader2CNorm.BytesRef()); IndexReader reader3C = (IndexReader) reader2C.Clone(); SegmentReader segmentReader3C = SegmentReader.GetOnlySegmentReader(reader3C); Norm reader3CCNorm = (Norm) segmentReader3C.norms_ForNUnit["field1"]; Assert.AreEqual(3, reader3CCNorm.BytesRef().RefCount()); // edit a norm and the refcount should be 1 IndexReader reader4C = (IndexReader) reader3C.Clone(); SegmentReader segmentReader4C = SegmentReader.GetOnlySegmentReader(reader4C); Assert.AreEqual(4, reader3CCNorm.BytesRef().RefCount()); reader4C.SetNorm(5, "field1", 0.33f); // generate a cannot update exception in reader1 try { reader3C.SetNorm(1, "field1", 0.99f); Assert.Fail("did not hit expected exception"); } catch (System.Exception ex) { // expected } // norm values should be different Assert.IsTrue(Similarity.DecodeNorm(segmentReader3C.Norms("field1")[5]) != Similarity.DecodeNorm(segmentReader4C.Norms("field1")[5])); Norm reader4CCNorm = (Norm) segmentReader4C.norms_ForNUnit["field1"]; Assert.AreEqual(3, reader3CCNorm.BytesRef().RefCount()); Assert.AreEqual(1, reader4CCNorm.BytesRef().RefCount()); IndexReader reader5C = (IndexReader) reader4C.Clone(); SegmentReader segmentReader5C = SegmentReader.GetOnlySegmentReader(reader5C); Norm reader5CCNorm = (Norm) segmentReader5C.norms_ForNUnit["field1"]; reader5C.SetNorm(5, "field1", 0.7f); Assert.AreEqual(1, reader5CCNorm.BytesRef().RefCount()); reader5C.Close(); reader4C.Close(); reader3C.Close(); reader2C.Close(); reader1.Close(); dir1.Close(); } private void CreateIndex(Directory dir) { IndexWriter iw = new IndexWriter(dir, anlzr, true, IndexWriter.MaxFieldLength.LIMITED); iw.SetMaxBufferedDocs(5); iw.SetMergeFactor(3); iw.SetSimilarity(similarityOne); iw.SetUseCompoundFile(true); iw.Close(); } private void ModifyNormsForF1(Directory dir) { IndexReader ir = IndexReader.Open(dir); ModifyNormsForF1(ir); } private void ModifyNormsForF1(IndexReader ir) { int n = ir.MaxDoc(); // System.out.println("modifyNormsForF1 maxDoc: "+n); for (int i = 0; i < n; i += 3) { // modify for every third doc int k = (i * 3) % modifiedNorms.Count; float origNorm = (float) ((System.Single) modifiedNorms[i]); float newNorm = (float) ((System.Single) modifiedNorms[k]); // System.out.println("Modifying: for "+i+" from "+origNorm+" to // "+newNorm); // System.out.println(" and: for "+k+" from "+newNorm+" to "+origNorm); modifiedNorms[i] = (float) newNorm; modifiedNorms[k] = (float) origNorm; ir.SetNorm(i, "f" + 1, newNorm); ir.SetNorm(k, "f" + 1, origNorm); // System.out.println("setNorm i: "+i); // break; } // ir.close(); } private void VerifyIndex(Directory dir) { IndexReader ir = IndexReader.Open(dir); VerifyIndex(ir); ir.Close(); } private void VerifyIndex(IndexReader ir) { for (int i = 0; i < NUM_FIELDS; i++) { System.String field = "f" + i; byte[] b = ir.Norms(field); Assert.AreEqual(numDocNorms, b.Length, "number of norms mismatches"); System.Collections.ArrayList storedNorms = (i == 1?modifiedNorms:norms); for (int j = 0; j < b.Length; j++) { float norm = Similarity.DecodeNorm(b[j]); float norm1 = (float) ((System.Single) storedNorms[j]); Assert.AreEqual(norm, norm1, 0.000001, "stored norm value of " + field + " for doc " + j + " is " + norm + " - a mismatch!"); } } } private void AddDocs(Directory dir, int ndocs, bool compound) { IndexWriter iw = new IndexWriter(dir, anlzr, false, IndexWriter.MaxFieldLength.LIMITED); iw.SetMaxBufferedDocs(5); iw.SetMergeFactor(3); iw.SetSimilarity(similarityOne); iw.SetUseCompoundFile(compound); for (int i = 0; i < ndocs; i++) { iw.AddDocument(NewDoc()); } iw.Close(); } // create the next document private Document NewDoc() { Document d = new Document(); float boost = NextNorm(); for (int i = 0; i < 10; i++) { Field f = new Field("f" + i, "v" + i, Field.Store.NO, Field.Index.NOT_ANALYZED); f.SetBoost(boost); d.Add(f); } return d; } // return unique norm values that are unchanged by encoding/decoding private float NextNorm() { float norm = lastNorm + normDelta; do { float norm1 = Similarity.DecodeNorm(Similarity.EncodeNorm(norm)); if (norm1 > lastNorm) { // System.out.println(norm1+" > "+lastNorm); norm = norm1; break; } norm += normDelta; } while (true); norms.Insert(numDocNorms, (float) norm); modifiedNorms.Insert(numDocNorms, (float) norm); // System.out.println("creating norm("+numDocNorms+"): "+norm); numDocNorms++; lastNorm = (norm > 10?0:norm); // there's a limit to how many distinct // values can be stored in a ingle byte return norm; } } }