Package org.hibernate.search.query

Source Code of org.hibernate.search.query.FullTextQueryImpl

//$Id: FullTextQueryImpl.java 15387 2008-10-24 15:04:07Z hardy.ferentschik $
package org.hibernate.search.query;

import java.io.IOException;
import java.lang.reflect.InvocationTargetException;
import java.util.ArrayList;
import java.util.Collections;
import java.util.HashMap;
import java.util.HashSet;
import java.util.Iterator;
import java.util.List;
import java.util.Map;
import java.util.Set;

import org.apache.lucene.index.IndexReader;
import org.apache.lucene.index.Term;
import org.apache.lucene.search.BooleanClause;
import org.apache.lucene.search.BooleanQuery;
import org.apache.lucene.search.Explanation;
import org.apache.lucene.search.Filter;
import org.apache.lucene.search.Hits;
import org.apache.lucene.search.IndexSearcher;
import org.apache.lucene.search.Searcher;
import org.apache.lucene.search.Similarity;
import org.apache.lucene.search.Sort;
import org.apache.lucene.search.TermQuery;
import org.slf4j.Logger;

import org.hibernate.Criteria;
import org.hibernate.HibernateException;
import org.hibernate.LockMode;
import org.hibernate.Query;
import org.hibernate.ScrollMode;
import org.hibernate.ScrollableResults;
import org.hibernate.Session;
import org.hibernate.annotations.common.util.ReflectHelper;
import org.hibernate.engine.SessionImplementor;
import org.hibernate.engine.query.ParameterMetadata;
import org.hibernate.impl.AbstractQueryImpl;
import org.hibernate.impl.CriteriaImpl;
import org.hibernate.search.FullTextFilter;
import org.hibernate.search.FullTextQuery;
import org.hibernate.search.SearchException;
import org.hibernate.search.engine.DocumentBuilder;
import org.hibernate.search.engine.DocumentExtractor;
import org.hibernate.search.engine.EntityInfo;
import org.hibernate.search.engine.FilterDef;
import org.hibernate.search.engine.Loader;
import org.hibernate.search.engine.MultiClassesQueryLoader;
import org.hibernate.search.engine.ProjectionLoader;
import org.hibernate.search.engine.QueryLoader;
import org.hibernate.search.engine.SearchFactoryImplementor;
import org.hibernate.search.filter.ChainedFilter;
import org.hibernate.search.filter.FilterKey;
import org.hibernate.search.filter.StandardFilterKey;
import org.hibernate.search.reader.ReaderProvider;
import static org.hibernate.search.reader.ReaderProviderHelper.getIndexReaders;
import org.hibernate.search.store.DirectoryProvider;
import org.hibernate.search.util.ContextHelper;
import static org.hibernate.search.util.FilterCacheModeTypeHelper.cacheInstance;
import static org.hibernate.search.util.FilterCacheModeTypeHelper.cacheResults;
import org.hibernate.search.util.LoggerFactory;
import org.hibernate.transform.ResultTransformer;

/**
* Implementation of {@link org.hibernate.search.FullTextQuery}
*
* @author Emmanuel Bernard
* @author Hardy Ferentschik
*/
//TODO implements setParameter()
public class FullTextQueryImpl extends AbstractQueryImpl implements FullTextQuery {
  private static final Logger log = LoggerFactory.make();
  private final org.apache.lucene.search.Query luceneQuery;
  private Class<?>[] classes;
  private Set<Class<?>> classesAndSubclasses;
  //optimization: if we can avoid the filter clause (we can most of the time) do it as it has a significant perf impact
  private boolean needClassFilterClause;
  private Integer firstResult;
  private Integer maxResults;
  private Integer resultSize;
  private Sort sort;
  private Filter filter;
  private Criteria criteria;
  private String[] indexProjection;
  private ResultTransformer resultTransformer;
  private SearchFactoryImplementor searchFactoryImplementor;
  private Map<String, FullTextFilterImpl> filterDefinitions;
  private int fetchSize = 1;

  /**
   * classes must be immutable
   */
  public FullTextQueryImpl(org.apache.lucene.search.Query query, Class[] classes, SessionImplementor session,
               ParameterMetadata parameterMetadata) {
    //TODO handle flushMode
    super( query.toString(), null, session, parameterMetadata );
    this.luceneQuery = query;
    this.classes = classes;
  }

  /**
   * {@inheritDoc}
   */
  public FullTextQuery setSort(Sort sort) {
    this.sort = sort;
    return this;
  }

  /**
   * {@inheritDoc}
   */
  public FullTextQuery setFilter(Filter filter) {
    this.filter = filter;
    return this;
  }

  /**
   * Return an interator on the results.
   * Retrieve the object one by one (initialize it during the next() operation)
   */
  public Iterator iterate() throws HibernateException {
    //implement an interator which keep the id/class for each hit and get the object on demand
    //cause I can't keep the searcher and hence the hit opened. I dont have any hook to know when the
    //user stop using it
    //scrollable is better in this area

    SearchFactoryImplementor searchFactoryImplementor = ContextHelper.getSearchFactoryBySFI( session );
    //find the directories
    IndexSearcher searcher = buildSearcher( searchFactoryImplementor );
    if ( searcher == null ) {
      return new IteratorImpl( Collections.EMPTY_LIST, noLoader );
    }
    try {
      QueryAndHits queryAndHits = getQueryAndHits( searcher );
      int first = first();
      int max = max( first, queryAndHits.hits );
      Session sess = (Session) this.session;

      int size = max - first + 1 < 0 ? 0 : max - first + 1;
      List<EntityInfo> infos = new ArrayList<EntityInfo>( size );
      DocumentExtractor extractor = new DocumentExtractor( queryAndHits.preparedQuery, searcher, searchFactoryImplementor, indexProjection );
      for (int index = first; index <= max; index++) {
        //TODO use indexSearcher.getIndexReader().document( hits.id(index), FieldSelector(indexProjection) );
        infos.add( extractor.extract( queryAndHits.hits, index ) );
      }
      Loader loader = getLoader( sess, searchFactoryImplementor );
      return new IteratorImpl( infos, loader );
    }
    catch (IOException e) {
      throw new HibernateException( "Unable to query Lucene index", e );
    }
    finally {
      try {
        closeSearcher( searcher, searchFactoryImplementor.getReaderProvider() );
      }
      catch (SearchException e) {
        log.warn( "Unable to properly close searcher during lucene query: " + getQueryString(), e );
      }
    }
  }

  private Loader getLoader(Session session, SearchFactoryImplementor searchFactoryImplementor) {
    if ( indexProjection != null ) {
      ProjectionLoader loader = new ProjectionLoader();
      loader.init( session, searchFactoryImplementor, resultTransformer, indexProjection );
      return loader;
    }
    if ( criteria != null ) {
      if ( classes.length > 1 ) throw new SearchException( "Cannot mix criteria and multiple entity types" );
      if ( criteria instanceof CriteriaImpl ) {
        String targetEntity = ( (CriteriaImpl) criteria ).getEntityOrClassName();
        if ( classes.length == 1 && !classes[0].getName().equals( targetEntity ) ) {
          throw new SearchException( "Criteria query entity should match query entity" );
        }
        else {
          try {
            Class entityType = ReflectHelper.classForName( targetEntity );
            classes = new Class[] { entityType };
          }
          catch (ClassNotFoundException e) {
            throw new SearchException( "Unable to load entity class from criteria: " + targetEntity, e );
          }
        }
      }
      QueryLoader loader = new QueryLoader();
      loader.init( session, searchFactoryImplementor );
      loader.setEntityType( classes[0] );
      loader.setCriteria( criteria );
      return loader;
    }
    else if ( classes.length == 1 ) {
      final QueryLoader loader = new QueryLoader();
      loader.init( session, searchFactoryImplementor );
      loader.setEntityType( classes[0] );
      return loader;
    }
    else {
      final MultiClassesQueryLoader loader = new MultiClassesQueryLoader();
      loader.init( session, searchFactoryImplementor );
      loader.setEntityTypes( classes );
      return loader;
    }
  }

  public ScrollableResults scroll() throws HibernateException {
    //keep the searcher open until the resultset is closed
    SearchFactoryImplementor searchFactory = ContextHelper.getSearchFactoryBySFI( session );

    //find the directories
    IndexSearcher searcher = buildSearcher( searchFactory );
    //FIXME: handle null searcher
    try {
      QueryAndHits queryAndHits = getQueryAndHits( searcher );
      int first = first();
      int max = max( first, queryAndHits.hits );
      DocumentExtractor extractor = new DocumentExtractor( queryAndHits.preparedQuery, searcher, searchFactory, indexProjection );
      Loader loader = getLoader( (Session) this.session, searchFactory );
      return new ScrollableResultsImpl( searcher, queryAndHits.hits, first, max, fetchSize, extractor, loader, searchFactory );
    }
    catch (IOException e) {
      //close only in case of exception
      try {
        closeSearcher( searcher, searchFactory.getReaderProvider() );
      }
      catch (SearchException ee) {
        //we have the initial issue already
      }
      throw new HibernateException( "Unable to query Lucene index", e );
    }
  }

  public ScrollableResults scroll(ScrollMode scrollMode) throws HibernateException {
    //TODO think about this scrollmode
    return scroll();
  }

  public List list() throws HibernateException {
    SearchFactoryImplementor searchFactoryImplementor = ContextHelper.getSearchFactoryBySFI( session );
    //find the directories
    IndexSearcher searcher = buildSearcher( searchFactoryImplementor );
    if ( searcher == null ) return Collections.EMPTY_LIST;
    try {
      QueryAndHits queryAndHits = getQueryAndHits( searcher );
      int first = first();
      int max = max( first, queryAndHits.hits );
      Session sess = (Session) this.session;

      int size = max - first + 1 < 0 ? 0 : max - first + 1;
      List<EntityInfo> infos = new ArrayList<EntityInfo>( size );
      DocumentExtractor extractor = new DocumentExtractor( queryAndHits.preparedQuery, searcher, searchFactoryImplementor, indexProjection );
      for (int index = first; index <= max; index++) {
        infos.add( extractor.extract( queryAndHits.hits, index ) );
      }
      Loader loader = getLoader( sess, searchFactoryImplementor );
      List list = loader.load( infos.toArray( new EntityInfo[infos.size()] ) );
      if ( resultTransformer == null || loader instanceof ProjectionLoader ) {
        //stay consistent with transformTuple which can only be executed during a projection
        return list;
      }
      else {
        return resultTransformer.transformList( list );
      }
    }
    catch (IOException e) {
      throw new HibernateException( "Unable to query Lucene index", e );
    }
    finally {
      try {
        closeSearcher( searcher, searchFactoryImplementor.getReaderProvider() );
      }
      catch (SearchException e) {
        log.warn( "Unable to properly close searcher during lucene query: " + getQueryString(), e );
      }
    }
  }

  public Explanation explain(int documentId) {
    Explanation explanation = null;
    SearchFactoryImplementor searchFactoryImplementor = getSearchFactoryImplementor();
    Searcher searcher = buildSearcher( searchFactoryImplementor );
    if (searcher == null) {
      throw new SearchException("Unable to build explanation for document id:"
          + documentId + ". no index found");
    }
    try {
      org.apache.lucene.search.Query query = filterQueryByClasses( luceneQuery );
      buildFilters();
      explanation = searcher.explain( query, documentId );
    }
    catch (IOException e) {
      throw new HibernateException( "Unable to query Lucene index and build explanation", e );
    }
    finally {
      //searcher cannot be null
      try {
        closeSearcher( searcher, searchFactoryImplementor.getReaderProvider() );
      }
      catch (SearchException e) {
        log.warn( "Unable to properly close searcher during lucene query: " + getQueryString(), e );
      }
    }
    return explanation;
  }

  /**
   * Execute the lucene search and return the machting hits.
   *
   * @param searcher The index searcher.
   * @return The lucene hits.
   * @throws IOException in case there is an error executing the lucene search.
   */
  private QueryAndHits getQueryAndHits(Searcher searcher) throws IOException {
    Hits hits;
    org.apache.lucene.search.Query query = filterQueryByClasses( luceneQuery );
    buildFilters();
    hits = searcher.search( query, filter, sort );
    setResultSize( hits );
    return new QueryAndHits( query, hits );
  }

  private void buildFilters() {
    if ( filterDefinitions == null || filterDefinitions.size() == 0 ) {
      return; // there is nothing to do if we don't have any filter definitions
    }

    ChainedFilter chainedFilter = new ChainedFilter();
    for (FullTextFilterImpl fullTextFilter : filterDefinitions.values()) {
      Filter filter = buildLuceneFilter(fullTextFilter);
      chainedFilter.addFilter( filter );
    }

    if ( filter != null ) chainedFilter.addFilter( filter );
    filter = chainedFilter;
  }

  /**
   * Builds a Lucene filter using the given <code>FullTextFilter</code>.
   *
   * @param fullTextFilter the Hibernate specific <code>FullTextFilter</code> used to create the
   * Lucene <code>Filter</code>.
   * @return the Lucene filter mapped to the filter definition
   */
  private Filter buildLuceneFilter(FullTextFilterImpl fullTextFilter) {

    SearchFactoryImplementor searchFactoryImplementor = getSearchFactoryImplementor();

    /*
     * FilterKey implementations and Filter(Factory) do not have to be threadsafe wrt their parameter injection
     * as FilterCachingStrategy ensure a memory barrier between concurrent thread calls
     */
    FilterDef def = searchFactoryImplementor.getFilterDefinition( fullTextFilter.getName() );
    Object instance = createFilterInstance(fullTextFilter, def);
    FilterKey key = createFilterKey(def, instance);

    // try to get the filter out of the cache
    Filter filter = cacheInstance( def.getCacheMode() ) ?
        searchFactoryImplementor.getFilterCachingStrategy().getCachedFilter( key ) :
        null;

    if ( filter == null ) {
      filter = createFilter(def, instance);

      // add filter to cache if we have to
      if ( cacheInstance( def.getCacheMode() ) ) {
        searchFactoryImplementor.getFilterCachingStrategy().addCachedFilter( key, filter );
      }
    }
    return filter;
  }

  private Filter createFilter(FilterDef def, Object instance) {
    Filter filter = null;
    if ( def.getFactoryMethod() != null ) {
      try {
        filter = (Filter) def.getFactoryMethod().invoke( instance );
      }
      catch (IllegalAccessException e) {
        throw new SearchException( "Unable to access @Factory method: "
            + def.getImpl().getName() + "." + def.getFactoryMethod().getName() );
      }
      catch (InvocationTargetException e) {
        throw new SearchException( "Unable to access @Factory method: "
            + def.getImpl().getName() + "." + def.getFactoryMethod().getName() );
      }
      catch (ClassCastException e) {
        throw new SearchException( "@Key method does not return a org.apache.lucene.search.Filter class: "
            + def.getImpl().getName() + "." + def.getFactoryMethod().getName() );
      }
    }
    else {
      try {
        filter = (Filter) instance;
      }
      catch (ClassCastException e) {
        throw new SearchException( "Filter implementation does not implement the Filter interface: "
            + def.getImpl().getName() + ". "
            + (def.getFactoryMethod() != null ? def.getFactoryMethod().getName() : ""), e );
      }
    }

    filter = addCachingWrapperFilter(filter, def);
    return filter;
  }

  /**
   * Decides whether to wrap the given filter around a <code>CachingWrapperFilter<code>.
   *
   * @param filter the filter which maybe gets wrapped.
   * @param def The filter definition used to decide whether wrapping should occur or not.
   * @return The original filter or wrapped filter depending on the information extracted from
   * <code>def</code>.
   */
  private Filter addCachingWrapperFilter(Filter filter, FilterDef def) {
    if ( cacheResults( def.getCacheMode() ) ) {
      int cachingWrapperFilterSize = getSearchFactoryImplementor().getFilterCacheBitResultsSize();
      filter = new org.hibernate.search.filter.CachingWrapperFilter(filter, cachingWrapperFilterSize);
    }

    return filter;
  }

  private FilterKey createFilterKey(FilterDef def, Object instance) {
    FilterKey key = null;
    if ( !cacheInstance( def.getCacheMode() ) ) {
      return key; // if the filter is not cached there is no key!
    }

    if ( def.getKeyMethod() == null ) {
      key = new FilterKey() {
        public int hashCode() {
          return getImpl().hashCode();
        }

        public boolean equals(Object obj) {
          if ( !( obj instanceof FilterKey ) ) return false;
          FilterKey that = (FilterKey) obj;
          return this.getImpl().equals( that.getImpl() );
        }
      };
    }
    else {
      try {
        key = (FilterKey) def.getKeyMethod().invoke( instance );
      }
      catch (IllegalAccessException e) {
        throw new SearchException( "Unable to access @Key method: "
            + def.getImpl().getName() + "." + def.getKeyMethod().getName() );
      }
      catch (InvocationTargetException e) {
        throw new SearchException( "Unable to access @Key method: "
            + def.getImpl().getName() + "." + def.getKeyMethod().getName() );
      }
      catch (ClassCastException e) {
        throw new SearchException( "@Key method does not return FilterKey: "
            + def.getImpl().getName() + "." + def.getKeyMethod().getName() );
      }
    }
    key.setImpl( def.getImpl() );

    //Make sure Filters are isolated by filter def name
    StandardFilterKey wrapperKey = new StandardFilterKey();
    wrapperKey.addParameter( def.getName() );
    wrapperKey.addParameter( key );
    return wrapperKey;
  }

  private Object createFilterInstance(FullTextFilterImpl fullTextFilter,
      FilterDef def) {
    Object instance;
    try {
      instance = def.getImpl().newInstance();
    }
    catch (InstantiationException e) {
      throw new SearchException( "Unable to create @FullTextFilterDef: " + def.getImpl(), e );
    }
    catch (IllegalAccessException e) {
      throw new SearchException( "Unable to create @FullTextFilterDef: " + def.getImpl(), e );
    }
    for (Map.Entry<String, Object> entry : fullTextFilter.getParameters().entrySet()) {
      def.invoke( entry.getKey(), instance, entry.getValue() );
    }
    if ( cacheInstance( def.getCacheMode() ) && def.getKeyMethod() == null && fullTextFilter.getParameters().size() > 0 ) {
      throw new SearchException( "Filter with parameters and no @Key method: " + fullTextFilter.getName() );
    }
    return instance;
  }

  private org.apache.lucene.search.Query filterQueryByClasses(org.apache.lucene.search.Query luceneQuery) {
    if ( ! needClassFilterClause ) {
      return luceneQuery;
    }
    else {
      //A query filter is more practical than a manual class filtering post query (esp on scrollable resultsets)
      //it also probably minimise the memory footprint 
      BooleanQuery classFilter = new BooleanQuery();
      //annihilate the scoring impact of DocumentBuilder.CLASS_FIELDNAME
      classFilter.setBoost( 0 );
      for (Class clazz : classesAndSubclasses) {
        Term t = new Term( DocumentBuilder.CLASS_FIELDNAME, clazz.getName() );
        TermQuery termQuery = new TermQuery( t );
        classFilter.add( termQuery, BooleanClause.Occur.SHOULD );
      }
      BooleanQuery filteredQuery = new BooleanQuery();
      filteredQuery.add( luceneQuery, BooleanClause.Occur.MUST );
      filteredQuery.add( classFilter, BooleanClause.Occur.MUST );
      return filteredQuery;
    }
  }

  private int max(int first, Hits hits) {
    return maxResults == null ?
        hits.length() - 1 :
        maxResults + first < hits.length() ?
            first + maxResults - 1 :
            hits.length() - 1;
  }

  private int first() {
    return firstResult != null ?
        firstResult :
        0;
  }


  /**
   * can return null
   * TODO change classesAndSubclasses by side effect, which is a mismatch with the Searcher return, fix that.
   */
  private IndexSearcher buildSearcher(SearchFactoryImplementor searchFactoryImplementor) {
    Map<Class<?>, DocumentBuilder<?>> builders = searchFactoryImplementor.getDocumentBuilders();
    List<DirectoryProvider> directories = new ArrayList<DirectoryProvider>();

    Similarity searcherSimilarity = null;
    //TODO check if caching this work for the last n list of classes makes a perf boost
    if ( classes == null || classes.length == 0 ) {
      // empty classes array means search over all indexed enities,
      // but we have to make sure there is at least one
      if ( builders.isEmpty() ) {
        throw new HibernateException( "There are no mapped entities (don't forget to add @Indexed to at least one class)." );
      }

      for (DocumentBuilder builder : builders.values()) {
        searcherSimilarity = checkSimilarity( searcherSimilarity, builder );
        final DirectoryProvider[] directoryProviders = builder.getDirectoryProviderSelectionStrategy().getDirectoryProvidersForAllShards();
        populateDirectories( directories, directoryProviders, searchFactoryImplementor );
      }
      classesAndSubclasses = null;
    }
    else {
      Set<Class<?>> involvedClasses = new HashSet<Class<?>>( classes.length );
      Collections.addAll( involvedClasses, classes );
      for (Class<?> clazz : classes) {
        DocumentBuilder<?> builder = builders.get( clazz );
        if ( builder != null ) involvedClasses.addAll( builder.getMappedSubclasses() );
      }

      for (Class clazz : involvedClasses) {
        DocumentBuilder builder = builders.get( clazz );
        //TODO should we rather choose a polymorphic path and allow non mapped entities
        if ( builder == null )
          throw new HibernateException( "Not a mapped entity (don't forget to add @Indexed): " + clazz );

        final DirectoryProvider[] directoryProviders = builder.getDirectoryProviderSelectionStrategy().getDirectoryProvidersForAllShards();
        searcherSimilarity = checkSimilarity( searcherSimilarity, builder );
        populateDirectories( directories, directoryProviders, searchFactoryImplementor );
      }
      classesAndSubclasses = involvedClasses;
    }

    //compute optimization needClassFilterClause
    //if at least one DP contains one class that is not part of the targeted classesAndSubclasses we can't optimize
    if ( classesAndSubclasses != null) {
      for (DirectoryProvider dp : directories) {
        final Set<Class<?>> classesInDirectoryProvider = searchFactoryImplementor.getClassesInDirectoryProvider( dp );
        // if a DP contains only one class, we know for sure it's part of classesAndSubclasses
        if ( classesInDirectoryProvider.size() > 1 ) {
          //risk of needClassFilterClause
          for (Class clazz : classesInDirectoryProvider) {
            if ( ! classesAndSubclasses.contains( clazz ) ) {
              this.needClassFilterClause = true;
              break;
            }
          }
        }
        if ( this.needClassFilterClause ) break;
      }
    }

    //set up the searcher
    final DirectoryProvider[] directoryProviders = directories.toArray( new DirectoryProvider[directories.size()] );
    IndexSearcher is = new IndexSearcher( searchFactoryImplementor.getReaderProvider().openReader( directoryProviders ) );
    is.setSimilarity( searcherSimilarity );
    return is;
  }

  private void populateDirectories(List<DirectoryProvider> directories, DirectoryProvider[] directoryProviders,
                   SearchFactoryImplementor searchFactoryImplementor) {
    for (DirectoryProvider provider : directoryProviders) {
      if ( !directories.contains( provider ) ) {
        directories.add( provider );
      }
    }
  }

  private Similarity checkSimilarity(Similarity similarity, DocumentBuilder builder) {
    if ( similarity == null ) {
      similarity = builder.getSimilarity();
    }
    else if ( !similarity.getClass().equals( builder.getSimilarity().getClass() ) ) {
      throw new HibernateException( "Cannot perform search on two entities with differing Similarity implementations (" + similarity.getClass().getName() + " & " + builder.getSimilarity().getClass().getName() + ")" );
    }

    return similarity;
  }

  private void closeSearcher(Searcher searcher, ReaderProvider readerProvider) {
    Set<IndexReader> indexReaders = getIndexReaders( searcher );

    for (IndexReader indexReader : indexReaders) {
      readerProvider.closeReader( indexReader );
    }
  }

  private void setResultSize(Hits hits) {
    resultSize = hits.length();
  }


  public int getResultSize() {
    if ( resultSize == null ) {
      //get result size without object initialization
      SearchFactoryImplementor searchFactoryImplementor = ContextHelper.getSearchFactoryBySFI( session );
      IndexSearcher searcher = buildSearcher( searchFactoryImplementor );
      if ( searcher == null ) {
        resultSize = 0;
      }
      else {
        Hits hits;
        try {
          hits = getQueryAndHits( searcher ).hits;
          resultSize = hits.length();
        }
        catch (IOException e) {
          throw new HibernateException( "Unable to query Lucene index", e );
        }
        finally {
          //searcher cannot be null
          try {
            closeSearcher( searcher, searchFactoryImplementor.getReaderProvider() );
            //searchFactoryImplementor.getReaderProvider().closeReader( searcher.getIndexReader() );
          }
          catch (SearchException e) {
            log.warn( "Unable to properly close searcher during lucene query: " + getQueryString(), e );
          }
        }
      }
    }
    return this.resultSize;
  }

  public FullTextQuery setCriteriaQuery(Criteria criteria) {
    this.criteria = criteria;
    return this;
  }

  public FullTextQuery setProjection(String... fields) {
    if ( fields == null || fields.length == 0 ) {
      this.indexProjection = null;
    }
    else {
      this.indexProjection = fields;
    }
    return this;
  }

  public FullTextQuery setFirstResult(int firstResult) {
    if ( firstResult < 0 ) {
      throw new IllegalArgumentException( "'first' pagination parameter less than 0" );
    }
    this.firstResult = firstResult;
    return this;
  }

  public FullTextQuery setMaxResults(int maxResults) {
    if ( maxResults < 0 ) {
      throw new IllegalArgumentException( "'max' pagination parameter less than 0" );
    }
    this.maxResults = maxResults;
    return this;
  }

  public FullTextQuery setFetchSize(int fetchSize) {
    super.setFetchSize( fetchSize );
    if ( fetchSize <= 0 ) {
      throw new IllegalArgumentException( "'fetch size' parameter less than or equals to 0" );
    }
    this.fetchSize = fetchSize;
    return this;
  }

  @Override
  public FullTextQuery setResultTransformer(ResultTransformer transformer) {
    super.setResultTransformer( transformer );
    this.resultTransformer = transformer;
    return this;
  }

  public int executeUpdate() throws HibernateException {
    throw new HibernateException( "Not supported operation" );
  }

  public Query setLockMode(String alias, LockMode lockMode) {
    return null;
  }

  protected Map getLockModes() {
    return null;
  }

  public FullTextFilter enableFullTextFilter(String name) {
    if ( filterDefinitions == null ) {
      filterDefinitions = new HashMap<String, FullTextFilterImpl>();
    }
    FullTextFilterImpl filterDefinition = filterDefinitions.get( name );
    if ( filterDefinition != null ) return filterDefinition;

    filterDefinition = new FullTextFilterImpl();
    filterDefinition.setName( name );
    FilterDef filterDef = getSearchFactoryImplementor().getFilterDefinition( name );
    if ( filterDef == null ) {
      throw new SearchException( "Unkown @FullTextFilter: " + name );
    }
    filterDefinitions.put( name, filterDefinition );
    return filterDefinition;
  }

  public void disableFullTextFilter(String name) {
    filterDefinitions.remove( name );
  }

  private SearchFactoryImplementor getSearchFactoryImplementor() {
    if ( searchFactoryImplementor == null ) {
      searchFactoryImplementor = ContextHelper.getSearchFactoryBySFI( session );
    }
    return searchFactoryImplementor;
  }

  private static Loader noLoader = new Loader() {
    public void init(Session session, SearchFactoryImplementor searchFactoryImplementor) {
    }

    public Object load(EntityInfo entityInfo) {
      throw new UnsupportedOperationException( "noLoader should not be used" );
    }

    public List load(EntityInfo... entityInfos) {
      throw new UnsupportedOperationException( "noLoader should not be used" );
    }
  };

  private static class QueryAndHits {
    private QueryAndHits(org.apache.lucene.search.Query preparedQuery, Hits hits) {
      this.preparedQuery = preparedQuery;
      this.hits = hits;
    }

    public final org.apache.lucene.search.Query preparedQuery;
    public final Hits hits;
  }
}
TOP

Related Classes of org.hibernate.search.query.FullTextQueryImpl

TOP
Copyright © 2018 www.massapi.com. All rights reserved.
All source code are property of their respective owners. Java is a trademark of Sun Microsystems, Inc and owned by ORACLE Inc. Contact coftware#gmail.com.