// $Id: AnalyzerTest.java 15644 2008-12-03 09:54:33Z hardy.ferentschik $
package org.hibernate.search.test.analyzer;
import java.util.HashSet;
import java.util.Set;
import javax.print.attribute.HashAttributeSet;
import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.analysis.Token;
import org.apache.lucene.analysis.standard.StandardAnalyzer;
import org.apache.lucene.queryParser.QueryParser;
import org.slf4j.Logger;
import org.hibernate.Transaction;
import org.hibernate.annotations.common.reflection.ReflectionManager;
import org.hibernate.annotations.common.reflection.XClass;
import org.hibernate.search.FullTextQuery;
import org.hibernate.search.FullTextSession;
import org.hibernate.search.Search;
import org.hibernate.search.SearchFactory;
import org.hibernate.search.SearchException;
import org.hibernate.search.impl.InitContext;
import org.hibernate.search.cfg.SearchConfiguration;
import org.hibernate.search.cfg.SearchConfigurationFromHibernateCore;
import org.hibernate.search.engine.DocumentBuilderContainedEntity;
import org.hibernate.search.test.SearchTestCase;
import org.hibernate.search.test.util.AnalyzerUtils;
import org.hibernate.search.util.LoggerFactory;
/**
* @author Emmanuel Bernard
* @author Hardy Ferentschik
*/
public class AnalyzerTest extends SearchTestCase {
public static final Logger log = LoggerFactory.make();
public void testAnalyzerDiscriminator() throws Exception {
Article germanArticle = new Article();
germanArticle.setLanguage( "de" );
germanArticle.setText( "aufeinanderschl\u00FCgen" );
Set<Article> references = new HashSet<Article>();
references.add( germanArticle );
Article englishArticle = new Article();
englishArticle.setLanguage( "en" );
englishArticle.setText( "acknowledgment" );
englishArticle.setReferences( references );
FullTextSession s = Search.getFullTextSession( openSession() );
Transaction tx = s.beginTransaction();
s.persist( englishArticle );
tx.commit();
tx = s.beginTransaction();
// at query time we use a standard analyzer. We explicitly search for tokens which can only be found if the
// right language specific stemmer was used at index time
QueryParser parser = new QueryParser( "references.text", new StandardAnalyzer() );
org.apache.lucene.search.Query luceneQuery = parser.parse( "aufeinanderschlug" );
FullTextQuery query = s.createFullTextQuery( luceneQuery );
assertEquals( 1, query.getResultSize() );
parser = new QueryParser( "text", new StandardAnalyzer() );
luceneQuery = parser.parse( "acknowledg" );
query = s.createFullTextQuery( luceneQuery );
assertEquals( 1, query.getResultSize() );
tx.commit();
s.close();
}
public void testMultipleAnalyzerDiscriminatorDefinitions() throws Exception {
SearchConfigurationFromHibernateCore searchConfig = new SearchConfigurationFromHibernateCore( cfg );
ReflectionManager reflectionManager = searchConfig.getReflectionManager();
XClass xclass = reflectionManager.toXClass( BlogEntry.class );
InitContext context = new InitContext( searchConfig );
try {
new DocumentBuilderContainedEntity( xclass, context, reflectionManager );
fail();
}
catch ( SearchException e ) {
assertTrue( "Wrong error message", e.getMessage().startsWith( "Multiple AnalyzerDiscriminator defined in the same class hierarchy" ));
}
}
public void testScopedAnalyzers() throws Exception {
MyEntity en = new MyEntity();
en.setEntity( "Entity" );
en.setField( "Field" );
en.setProperty( "Property" );
en.setComponent( new MyComponent() );
en.getComponent().setComponentProperty( "component property" );
FullTextSession s = Search.getFullTextSession( openSession() );
Transaction tx = s.beginTransaction();
s.persist( en );
tx.commit();
tx = s.beginTransaction();
QueryParser parser = new QueryParser( "id", new StandardAnalyzer() );
org.apache.lucene.search.Query luceneQuery = parser.parse( "entity:alarm" );
FullTextQuery query = s.createFullTextQuery( luceneQuery, MyEntity.class );
assertEquals( 1, query.getResultSize() );
luceneQuery = parser.parse( "property:cat" );
query = s.createFullTextQuery( luceneQuery, MyEntity.class );
assertEquals( 1, query.getResultSize() );
luceneQuery = parser.parse( "field:energy" );
query = s.createFullTextQuery( luceneQuery, MyEntity.class );
assertEquals( 1, query.getResultSize() );
luceneQuery = parser.parse( "component.componentProperty:noise" );
query = s.createFullTextQuery( luceneQuery, MyEntity.class );
assertEquals( 1, query.getResultSize() );
s.delete( query.uniqueResult() );
tx.commit();
s.close();
}
public void testScopedAnalyzersFromSearchFactory() throws Exception {
FullTextSession session = Search.getFullTextSession( openSession() );
SearchFactory searchFactory = session.getSearchFactory();
Analyzer analyzer = searchFactory.getAnalyzer( MyEntity.class );
// you can pass what so ever into the analysis since the used analyzers are
// returning the same tokens all the time. We just want to make sure that
// the right analyzers are used.
Token[] tokens = AnalyzerUtils.tokensFromAnalysis( analyzer, "entity", "" );
AnalyzerUtils.assertTokensEqual( tokens, new String[] { "alarm", "dog", "performance" } );
tokens = AnalyzerUtils.tokensFromAnalysis( analyzer, "property", "" );
AnalyzerUtils.assertTokensEqual( tokens, new String[] { "sound", "cat", "speed" } );
tokens = AnalyzerUtils.tokensFromAnalysis( analyzer, "field", "" );
AnalyzerUtils.assertTokensEqual( tokens, new String[] { "music", "elephant", "energy" } );
tokens = AnalyzerUtils.tokensFromAnalysis( analyzer, "component.componentProperty", "" );
AnalyzerUtils.assertTokensEqual( tokens, new String[] { "noise", "mouse", "light" } );
// test border cases
try {
searchFactory.getAnalyzer( ( Class ) null );
}
catch ( IllegalArgumentException iae ) {
log.debug( "success" );
}
try {
searchFactory.getAnalyzer( String.class );
}
catch ( IllegalArgumentException iae ) {
log.debug( "success" );
}
session.close();
}
protected Class[] getMappings() {
return new Class[] { MyEntity.class, Article.class };
}
}