package org.apache.lucene.queryparser.analyzing; /* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import java.io.IOException; import java.io.Reader; import java.util.Map; import java.util.TreeMap; import org.apache.lucene.analysis.Analyzer; import org.apache.lucene.analysis.MockAnalyzer; import org.apache.lucene.analysis.MockBytesAnalyzer; import org.apache.lucene.analysis.MockTokenFilter; import org.apache.lucene.analysis.MockTokenizer; import org.apache.lucene.analysis.TokenFilter; import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.Tokenizer; import org.apache.lucene.analysis.tokenattributes.CharTermAttribute; import org.apache.lucene.document.Document; import org.apache.lucene.document.Field; import org.apache.lucene.document.FieldType; import org.apache.lucene.index.DirectoryReader; import org.apache.lucene.index.RandomIndexWriter; import org.apache.lucene.queryparser.classic.ParseException; import org.apache.lucene.queryparser.classic.QueryParser; import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.search.Query; import org.apache.lucene.store.Directory; import org.apache.lucene.util.LuceneTestCase; import org.apache.lucene.util.LuceneTestCase.SuppressCodecs; /** */ @SuppressCodecs("Lucene3x") // binary terms public class TestAnalyzingQueryParser extends LuceneTestCase { private final static String FIELD = "field"; private Analyzer a; private String[] wildcardInput; private String[] wildcardExpected; private String[] prefixInput; private String[] prefixExpected; private String[] rangeInput; private String[] rangeExpected; private String[] fuzzyInput; private String[] fuzzyExpected; private Map<String, String> wildcardEscapeHits = new TreeMap<>(); private Map<String, String> wildcardEscapeMisses = new TreeMap<>(); @Override public void setUp() throws Exception { super.setUp(); wildcardInput = new String[] { "*bersetzung über*ung", "Mötley Cr\u00fce Mötl?* Crü?", "Renée Zellweger Ren?? Zellw?ger" }; wildcardExpected = new String[] { "*bersetzung uber*ung", "motley crue motl?* cru?", "renee zellweger ren?? zellw?ger" }; prefixInput = new String[] { "übersetzung übersetz*", "Mötley Crüe Mötl* crü*", "René? Zellw*" }; prefixExpected = new String[] { "ubersetzung ubersetz*", "motley crue motl* cru*", "rene? zellw*" }; rangeInput = new String[] { "[aa TO bb]", "{Anaïs TO Zoé}" }; rangeExpected = new String[] { "[aa TO bb]", "{anais TO zoe}" }; fuzzyInput = new String[] { "Übersetzung Übersetzung~0.9", "Mötley Crüe Mötley~0.75 Crüe~0.5", "Renée Zellweger Renée~0.9 Zellweger~" }; fuzzyExpected = new String[] { "ubersetzung ubersetzung~1", "motley crue motley~1 crue~2", "renee zellweger renee~0 zellweger~2" }; wildcardEscapeHits.put("mö*tley", "moatley"); // need to have at least one genuine wildcard to trigger the wildcard analysis // hence the * before the y wildcardEscapeHits.put("mö\\*tl*y", "mo*tley"); // escaped backslash then true wildcard wildcardEscapeHits.put("mö\\\\*tley", "mo\\atley"); // escaped wildcard then true wildcard wildcardEscapeHits.put("mö\\??ley", "mo?tley"); // the first is an escaped * which should yield a miss wildcardEscapeMisses.put("mö\\*tl*y", "moatley"); a = new ASCIIAnalyzer(); } public void testSingleChunkExceptions() { boolean ex = false; String termStr = "the*tre"; Analyzer stopsAnalyzer = new MockAnalyzer (random(), MockTokenizer.WHITESPACE, true, MockTokenFilter.ENGLISH_STOPSET); try { String q = parseWithAnalyzingQueryParser(termStr, stopsAnalyzer, true); } catch (ParseException e){ if (e.getMessage().contains("returned nothing")){ ex = true; } } assertEquals("Should have returned nothing", true, ex); ex = false; AnalyzingQueryParser qp = new AnalyzingQueryParser(TEST_VERSION_CURRENT, FIELD, a); try{ qp.analyzeSingleChunk(FIELD, "", "not a single chunk"); } catch (ParseException e){ if (e.getMessage().contains("multiple terms")){ ex = true; } } assertEquals("Should have produced multiple terms", true, ex); } public void testWildcardAlone() throws ParseException { //seems like crazy edge case, but can be useful in concordance boolean pex = false; try{ Query q = getAnalyzedQuery("*", a, false); } catch (ParseException e){ pex = true; } assertEquals("Wildcard alone with allowWildcard=false", true, pex); pex = false; try { String qString = parseWithAnalyzingQueryParser("*", a, true); assertEquals("Every word", "*", qString); } catch (ParseException e){ pex = true; } assertEquals("Wildcard alone with allowWildcard=true", false, pex); } public void testWildCardEscapes() throws ParseException, IOException { for (Map.Entry<String, String> entry : wildcardEscapeHits.entrySet()){ Query q = getAnalyzedQuery(entry.getKey(), a, false); assertEquals("WildcardEscapeHits: " + entry.getKey(), true, isAHit(q, entry.getValue(), a)); } for (Map.Entry<String, String> entry : wildcardEscapeMisses.entrySet()){ Query q = getAnalyzedQuery(entry.getKey(), a, false); assertEquals("WildcardEscapeMisses: " + entry.getKey(), false, isAHit(q, entry.getValue(), a)); } } public void testWildCardQueryNoLeadingAllowed() { boolean ex = false; try{ String q = parseWithAnalyzingQueryParser(wildcardInput[0], a, false); } catch (ParseException e){ ex = true; } assertEquals("Testing initial wildcard not allowed", true, ex); } public void testWildCardQuery() throws ParseException { for (int i = 0; i < wildcardInput.length; i++) { assertEquals("Testing wildcards with analyzer " + a.getClass() + ", input string: " + wildcardInput[i], wildcardExpected[i], parseWithAnalyzingQueryParser(wildcardInput[i], a, true)); } } public void testPrefixQuery() throws ParseException { for (int i = 0; i < prefixInput.length; i++) { assertEquals("Testing prefixes with analyzer " + a.getClass() + ", input string: " + prefixInput[i], prefixExpected[i], parseWithAnalyzingQueryParser(prefixInput[i], a, false)); } } public void testRangeQuery() throws ParseException { for (int i = 0; i < rangeInput.length; i++) { assertEquals("Testing ranges with analyzer " + a.getClass() + ", input string: " + rangeInput[i], rangeExpected[i], parseWithAnalyzingQueryParser(rangeInput[i], a, false)); } } public void testFuzzyQuery() throws ParseException { for (int i = 0; i < fuzzyInput.length; i++) { assertEquals("Testing fuzzys with analyzer " + a.getClass() + ", input string: " + fuzzyInput[i], fuzzyExpected[i], parseWithAnalyzingQueryParser(fuzzyInput[i], a, false)); } } private String parseWithAnalyzingQueryParser(String s, Analyzer a, boolean allowLeadingWildcard) throws ParseException { Query q = getAnalyzedQuery(s, a, allowLeadingWildcard); return q.toString(FIELD); } private Query getAnalyzedQuery(String s, Analyzer a, boolean allowLeadingWildcard) throws ParseException { AnalyzingQueryParser qp = new AnalyzingQueryParser(TEST_VERSION_CURRENT, FIELD, a); qp.setAllowLeadingWildcard(allowLeadingWildcard); org.apache.lucene.search.Query q = qp.parse(s); return q; } final static class FoldingFilter extends TokenFilter { final CharTermAttribute termAtt = addAttribute(CharTermAttribute.class); public FoldingFilter(TokenStream input) { super(input); } @Override public boolean incrementToken() throws IOException { if (input.incrementToken()) { char term[] = termAtt.buffer(); for (int i = 0; i < term.length; i++) switch(term[i]) { case 'ü': term[i] = 'u'; break; case 'ö': term[i] = 'o'; break; case 'é': term[i] = 'e'; break; case 'ï': term[i] = 'i'; break; } return true; } else { return false; } } } final static class ASCIIAnalyzer extends Analyzer { @Override public TokenStreamComponents createComponents(String fieldName) { Tokenizer result = new MockTokenizer(MockTokenizer.WHITESPACE, true); return new TokenStreamComponents(result, new FoldingFilter(result)); } } // LUCENE-4176 public void testByteTerms() throws Exception { String s = "เข"; Analyzer analyzer = new MockBytesAnalyzer(); QueryParser qp = new AnalyzingQueryParser(TEST_VERSION_CURRENT, FIELD, analyzer); Query q = qp.parse("[เข TO เข]"); assertEquals(true, isAHit(q, s, analyzer)); } private boolean isAHit(Query q, String content, Analyzer analyzer) throws IOException{ Directory ramDir = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random(), ramDir, analyzer); Document doc = new Document(); FieldType fieldType = new FieldType(); fieldType.setIndexed(true); fieldType.setTokenized(true); fieldType.setStored(true); Field field = new Field(FIELD, content, fieldType); doc.add(field); writer.addDocument(doc); writer.close(); DirectoryReader ir = DirectoryReader.open(ramDir); IndexSearcher is = new IndexSearcher(ir); int hits = is.search(q, 10).totalHits; ir.close(); ramDir.close(); if (hits == 1){ return true; } else { return false; } } }