/* * Hibernate Search, full-text search for your domain model * * License: GNU Lesser General Public License (LGPL), version 2.1 or later * See the lgpl.txt file in the root directory or . */ package org.hibernate.search.engine.spi; import java.io.Serializable; import java.util.ArrayList; import java.util.Calendar; import java.util.Collection; import java.util.Collections; import java.util.Date; import java.util.HashMap; import java.util.HashSet; import java.util.List; import java.util.Map; import java.util.Set; import org.apache.lucene.analysis.Analyzer; import org.apache.lucene.document.DateTools; import org.apache.lucene.document.Document; import org.apache.lucene.document.DoubleDocValuesField; import org.apache.lucene.document.Field; import org.apache.lucene.document.Field.Index; import org.apache.lucene.document.FieldType; import org.apache.lucene.document.FloatDocValuesField; import org.apache.lucene.document.NumericDocValuesField; import org.apache.lucene.document.SortedDocValuesField; import org.apache.lucene.facet.sortedset.SortedSetDocValuesFacetField; import org.apache.lucene.index.IndexOptions; import org.apache.lucene.index.IndexableField; import org.apache.lucene.util.BytesRef; import org.hibernate.annotations.common.reflection.ReflectionManager; import org.hibernate.annotations.common.reflection.XClass; import org.hibernate.annotations.common.reflection.XMember; import org.hibernate.search.analyzer.Discriminator; import org.hibernate.search.annotations.ProvidedId; import org.hibernate.search.annotations.Store; import org.hibernate.search.backend.AddLuceneWork; import org.hibernate.search.backend.DeleteLuceneWork; import org.hibernate.search.backend.LuceneWork; import org.hibernate.search.backend.UpdateLuceneWork; import org.hibernate.search.bridge.FieldBridge; import org.hibernate.search.bridge.LuceneOptions; import org.hibernate.search.bridge.StringBridge; import org.hibernate.search.bridge.TwoWayFieldBridge; import org.hibernate.search.bridge.TwoWayStringBridge; import org.hibernate.search.bridge.builtin.NumericEncodingCalendarBridge; import org.hibernate.search.bridge.builtin.NumericEncodingDateBridge; import org.hibernate.search.bridge.builtin.NumericFieldBridge; import org.hibernate.search.bridge.builtin.StringEncodingDateBridge; import org.hibernate.search.bridge.builtin.impl.TwoWayString2FieldBridgeAdaptor; import org.hibernate.search.bridge.spi.ConversionContext; import org.hibernate.search.engine.ProjectionConstants; import org.hibernate.search.engine.impl.ConfigContext; import org.hibernate.search.engine.impl.FacetHandling; import org.hibernate.search.engine.impl.LuceneOptionsImpl; import org.hibernate.search.engine.metadata.impl.DocumentFieldMetadata; import org.hibernate.search.engine.metadata.impl.EmbeddedTypeMetadata; import org.hibernate.search.engine.metadata.impl.FacetMetadata; import org.hibernate.search.engine.metadata.impl.PropertyMetadata; import org.hibernate.search.engine.metadata.impl.SortableFieldMetadata; import org.hibernate.search.engine.metadata.impl.TypeMetadata; import org.hibernate.search.exception.AssertionFailure; import org.hibernate.search.exception.SearchException; import org.hibernate.search.spi.InstanceInitializer; import org.hibernate.search.util.impl.InternalAnalyzerUtils; import org.hibernate.search.util.impl.ReflectionHelper; import org.hibernate.search.util.logging.impl.Log; import org.hibernate.search.util.logging.impl.LoggerFactory; /** * Set up and provide a manager for classes which are directly annotated with {@code @Indexed}. * * @author Gavin King * @author Emmanuel Bernard * @author Sylvain Vieujot * @author Richard Hallier * @author Hardy Ferentschik */ public class DocumentBuilderIndexedEntity extends AbstractDocumentBuilder { /** * The tenant identifier. This is not a projection constant as we're not storing it. */ public static final String TENANT_ID_FIELDNAME = "__HSearch_TenantId"; private static final Log log = LoggerFactory.make(); private static final LuceneOptions NULL_EMBEDDED_MARKER_OPTIONS; static { DocumentFieldMetadata fieldMetadata = new DocumentFieldMetadata.Builder( null, Store.NO, Field.Index.NOT_ANALYZED_NO_NORMS, Field.TermVector.NO ) .boost( 1F ) .build(); NULL_EMBEDDED_MARKER_OPTIONS = new LuceneOptionsImpl( fieldMetadata, 1f, 1f ); } private static final FieldType TENANT_ID_FIELDTYPE = createTenantIdFieldType(); /** * Flag indicating whether {@link org.apache.lucene.search.IndexSearcher#doc(int, org.apache.lucene.index.StoredFieldVisitor)} * can be used in order to retrieve documents. This is only safe to do if we know that * all involved bridges are implementing TwoWayStringBridge. See HSEARCH-213. */ private boolean allowFieldSelectionInProjection = false; /** * Flag indicating whether there is an explicit id (@DocumentId or @Id) or not. When Search is used as make * for example using JBoss Cache Searchable the idKeywordName will be provided. */ private boolean idProvided; private final String identifierName; /** * The document field name of the document id */ private final String idFieldName; /** * The property metadata for the document id (not that in the case of a provided id the id getter can be {@code null}. */ private PropertyMetadata idPropertyMetadata; /** * Creates a document builder for entities annotated with @Indexed. * * @param clazz The class for which to build a DocumentBuilderContainedEntity * @param typeMetadata all the metadata for the entity type * @param context Handle to default configuration settings * @param reflectionManager Reflection manager to use for processing the annotations * @param optimizationBlackList mutable register, keeps track of types on which we need to disable collection events optimizations * @param instanceInitializer helper class for class object graph navigation */ public DocumentBuilderIndexedEntity(XClass clazz, TypeMetadata typeMetadata, ConfigContext context, ReflectionManager reflectionManager, Set optimizationBlackList, InstanceInitializer instanceInitializer) { super( clazz, typeMetadata, reflectionManager, optimizationBlackList, instanceInitializer ); ProvidedId providedIdAnnotation = findProvidedId( clazz, reflectionManager ); if ( providedIdAnnotation != null || context.isProvidedIdImplicit() ) { idProvided = true; } if ( idPropertyMetadata == null ) { idPropertyMetadata = getTypeMetadata().getIdPropertyMetadata(); } if ( idPropertyMetadata == null ) { throw log.noDocumentIdFoundException( clazz.getName() ); } idFieldName = idPropertyMetadata.getFieldMetadataSet().iterator().next().getName(); checkAllowFieldSelection(); if ( log.isDebugEnabled() ) { log.debugf( "Field selection in projections is set to %b for entity %s.", allowFieldSelectionInProjection, clazz ); } this.entityState = EntityState.INDEXED; this.identifierName = idProvided ? null : idPropertyMetadata.getPropertyAccessor().getName(); } public XMember getIdGetter() { return idPropertyMetadata.getPropertyAccessor(); } private ProvidedId findProvidedId(XClass clazz, ReflectionManager reflectionManager) { ProvidedId id = null; XClass currentClass = clazz; while ( id == null && ( !reflectionManager.equals( currentClass, Object.class ) ) ) { id = currentClass.getAnnotation( ProvidedId.class ); currentClass = currentClass.getSuperclass(); } return id; } @Override public void addWorkToQueue(String tenantId, Class entityClass, Object entity, Serializable id, boolean delete, boolean add, List queue, ConversionContext contextualBridge) { DocumentFieldMetadata idFieldMetadata = idPropertyMetadata.getFieldMetadata( idFieldName ); String idInString = objectToString( getIdBridge(), idFieldMetadata.getName(), id, contextualBridge ); if ( delete && !add ) { queue.add( new DeleteLuceneWork( tenantId, id, idInString, entityClass ) ); } else if ( add && !delete ) { queue.add( createAddWork( tenantId, entityClass, entity, id, idInString, getInstanceInitializer(), contextualBridge ) ); } else if ( add && delete ) { queue.add( createUpdateWork( tenantId, entityClass, entity, id, idInString, getInstanceInitializer(), contextualBridge ) ); } } private String objectToString(TwoWayFieldBridge bridge, String fieldName, Object value, ConversionContext conversionContext) { conversionContext.pushProperty( fieldName ); String stringValue; try { stringValue = conversionContext .setClass( getBeanClass() ) .twoWayConversionContext( bridge ) .objectToString( value ); } finally { conversionContext.popProperty(); } return stringValue; } private String objectToString(StringBridge bridge, String fieldName, Object value, ConversionContext conversionContext) { conversionContext.pushProperty( fieldName ); String stringValue; try { stringValue = conversionContext .setClass( getBeanClass() ) .stringConversionContext( bridge ) .objectToString( value ); } finally { conversionContext.popProperty(); } return stringValue; } public AddLuceneWork createAddWork(String tenantId, Class entityClass, Object entity, Serializable id, String idInString, InstanceInitializer sessionInitializer, ConversionContext conversionContext) { Map fieldToAnalyzerMap = new HashMap(); Document doc = getDocument( tenantId, entity, id, fieldToAnalyzerMap, sessionInitializer, conversionContext, null ); final AddLuceneWork addWork; if ( fieldToAnalyzerMap.isEmpty() ) { addWork = new AddLuceneWork( tenantId, id, idInString, entityClass, doc ); } else { addWork = new AddLuceneWork( tenantId, id, idInString, entityClass, doc, fieldToAnalyzerMap ); } return addWork; } public UpdateLuceneWork createUpdateWork(String tenantId, Class entityClass, Object entity, Serializable id, String idInString, InstanceInitializer sessionInitializer, ConversionContext contextualBridge) { Map fieldToAnalyzerMap = new HashMap(); Document doc = getDocument( tenantId, entity, id, fieldToAnalyzerMap, sessionInitializer, contextualBridge, null ); final UpdateLuceneWork addWork; if ( fieldToAnalyzerMap.isEmpty() ) { addWork = new UpdateLuceneWork( tenantId, id, idInString, entityClass, doc ); } else { addWork = new UpdateLuceneWork( tenantId, id, idInString, entityClass, doc, fieldToAnalyzerMap ); } return addWork; } /** * Builds the Lucene {@code Document} for a given entity instance and its id. * * @param tenantId the identifier of the tenant or null if there isn't one * @param instance The entity for which to build the matching Lucene {@code Document} * @param id the entity id. * @param fieldToAnalyzerMap this maps gets populated while generating the {@code Document}. * It allows to specify for any document field a named analyzer to use. This parameter cannot be {@code null}. * @param objectInitializer used to ensure that all objects are initialized * @param conversionContext a {@link org.hibernate.search.bridge.spi.ConversionContext} object. * @param includedFieldNames list of field names to consider. Others can be excluded. Null if all fields are considered. * * @return The Lucene {@code Document} for the specified entity. */ public Document getDocument( String tenantId, Object instance, Serializable id, Map fieldToAnalyzerMap, InstanceInitializer objectInitializer, ConversionContext conversionContext, String[] includedFieldNames) { // TODO as it is, includedFieldNames is not generally useful as we don't know if a field bridge creates specific fields or not // TODO only used at the moment to filter the id field and the class field if ( fieldToAnalyzerMap == null ) { throw new IllegalArgumentException( "fieldToAnalyzerMap cannot be null" ); } //sensible default for outside callers if ( objectInitializer == null ) { objectInitializer = getInstanceInitializer(); } Document doc = new Document(); FacetHandling faceting = new FacetHandling(); Class entityType = objectInitializer.getClass( instance ); float documentLevelBoost = getMetadata().getClassBoost( instance ); // add the class name of the entity to the document if ( containsFieldName( ProjectionConstants.OBJECT_CLASS, includedFieldNames ) ) { @SuppressWarnings( "deprecation" ) Field classField = new Field( ProjectionConstants.OBJECT_CLASS, entityType.getName(), Field.Store.YES, Field.Index.NOT_ANALYZED_NO_NORMS, Field.TermVector.NO ); doc.add( classField ); } addTenantIdIfRequired( tenantId, doc ); // now add the entity id to the document if ( containsFieldName( idFieldName, includedFieldNames ) ) { DocumentFieldMetadata idFieldMetaData = idPropertyMetadata.getFieldMetadata( idFieldName ); LuceneOptions luceneOptions = new LuceneOptionsImpl( idFieldMetaData, idFieldMetaData.getBoost(), documentLevelBoost ); final FieldBridge contextualizedBridge = conversionContext.oneWayConversionContext( getIdBridge() ); conversionContext.setClass( entityType ); conversionContext.pushProperty( idFieldMetaData.getName() ); try { contextualizedBridge.set( idFieldMetaData.getName(), id, doc, luceneOptions ); addSortFieldDocValues( doc, idPropertyMetadata, documentLevelBoost, id ); } finally { conversionContext.popProperty(); } } // finally add all other document fields Set processedFieldNames = new HashSet<>(); if (instance instanceof Collection) { System.out.println("TEST"); } buildDocumentFields( instance, doc, faceting, getMetadata(), fieldToAnalyzerMap, processedFieldNames, conversionContext, objectInitializer, documentLevelBoost, false ); doc = faceting.build( doc ); return doc; } private void addTenantIdIfRequired(String tenantId, Document doc) { if ( tenantId != null ) { Field tenantIdField = new Field( TENANT_ID_FIELDNAME, tenantId, TENANT_ID_FIELDTYPE ); doc.add( tenantIdField ); } } private static FieldType createTenantIdFieldType() { FieldType type = new FieldType(); type.setStored( false ); type.setOmitNorms( true ); type.setIndexOptions( IndexOptions.DOCS ); type.setTokenized( false ); type.setStoreTermVectorOffsets( false ); type.setStoreTermVectorPayloads( false ); type.setStoreTermVectorPositions( false ); type.setStoreTermVectors( false ); type.freeze(); return type; } /** * @param inheritedBoost Boost inherited from the parent structure of the given instance: the document-level boost * in case of a top-level field, the product of the document-level boost and the boost(s) of the parent * embeddable(s) in case of an embedded field */ private void buildDocumentFields(Object instance, Document doc, FacetHandling faceting, TypeMetadata typeMetadata, Map fieldToAnalyzerMap, Set processedFieldNames, ConversionContext conversionContext, InstanceInitializer objectInitializer, final float inheritedBoost, boolean multiValued) { // needed for field access: I cannot work in the proxied version Object unproxiedInstance = unproxy( instance, objectInitializer ); buildDocumentFieldForClassBridges( doc, typeMetadata, conversionContext, inheritedBoost, unproxiedInstance ); buildDocumentFieldsForProperties( doc, faceting, typeMetadata, conversionContext, objectInitializer, inheritedBoost, unproxiedInstance, multiValued ); // allow analyzer override for the fields added by the class and field bridges allowAnalyzerDiscriminatorOverride( doc, typeMetadata, fieldToAnalyzerMap, processedFieldNames, unproxiedInstance ); buildDocumentFieldsForEmbeddedObjects( doc, faceting, typeMetadata, fieldToAnalyzerMap, processedFieldNames, conversionContext, objectInitializer, inheritedBoost, unproxiedInstance, multiValued ); } private void buildDocumentFieldsForEmbeddedObjects(Document doc, FacetHandling faceting, TypeMetadata typeMetadata, Map fieldToAnalyzerMap, Set processedFieldNames, ConversionContext conversionContext, InstanceInitializer objectInitializer, float inheritedBoost, Object unproxiedInstance, boolean multiValued) { for ( EmbeddedTypeMetadata embeddedTypeMetadata : typeMetadata.getEmbeddedTypeMetadata() ) { XMember member = embeddedTypeMetadata.getEmbeddedGetter(); float embeddedBoost = inheritedBoost * embeddedTypeMetadata.getStaticBoost(); conversionContext.pushProperty( embeddedTypeMetadata.getEmbeddedFieldName() ); try { Object value = ReflectionHelper.getMemberValue( unproxiedInstance, member ); if ( value == null ) { processEmbeddedNullValue( doc, embeddedTypeMetadata, conversionContext ); continue; } switch ( embeddedTypeMetadata.getEmbeddedContainer() ) { case ARRAY: Object[] array = objectInitializer.initializeArray( (Object[]) value ); for ( Object arrayValue : array ) { buildDocumentFields( arrayValue, doc, faceting, embeddedTypeMetadata, fieldToAnalyzerMap, processedFieldNames, conversionContext, objectInitializer, embeddedBoost, true ); } break; case COLLECTION: Collection collection = objectInitializer.initializeCollection( (Collection) value ); for ( Object collectionValue : collection ) { buildDocumentFields( collectionValue, doc, faceting, embeddedTypeMetadata, fieldToAnalyzerMap, processedFieldNames, conversionContext, objectInitializer, embeddedBoost, true ); } break; case MAP: Map map = objectInitializer.initializeMap( (Map) value ); for ( Object collectionValue : map.values() ) { buildDocumentFields( collectionValue, doc, faceting, embeddedTypeMetadata, fieldToAnalyzerMap, processedFieldNames, conversionContext, objectInitializer, embeddedBoost, true ); } break; case OBJECT: buildDocumentFields( value, doc, faceting, embeddedTypeMetadata, fieldToAnalyzerMap, processedFieldNames, conversionContext, objectInitializer, embeddedBoost, multiValued ); break; default: throw new AssertionFailure( "Unknown embedded container: " + embeddedTypeMetadata.getEmbeddedContainer() ); } } finally { conversionContext.popProperty(); } } } /** * @param multiValued Whether the type whose properties should be added may appear more than once (within the same * role) in a document or not. That's the case if the type is (directly or indirectly) contained within an embedded * to-many association. */ private void buildDocumentFieldsForProperties(Document document, FacetHandling faceting, TypeMetadata typeMetadata, ConversionContext conversionContext, InstanceInitializer objectInitializer, float documentBoost, Object unproxiedInstance, boolean multiValued) { XMember previousMember = null; Object currentFieldValue = null; boolean isCollection = false; for ( PropertyMetadata propertyMetadata : typeMetadata.getAllPropertyMetadata() ) { XMember member = propertyMetadata.getPropertyAccessor(); if ( previousMember != member ) { currentFieldValue = unproxy( ReflectionHelper.getMemberValue( unproxiedInstance, member ), objectInitializer ); previousMember = member; if ( member.isCollection() ) { if ( currentFieldValue instanceof Collection ) { isCollection = true; objectInitializer.initializeCollection( (Collection) currentFieldValue ); } else if ( currentFieldValue instanceof Map ) { objectInitializer.initializeMap( (Map) currentFieldValue ); } } } try { conversionContext.pushProperty( propertyMetadata.getPropertyAccessorName() ); for ( DocumentFieldMetadata fieldMetadata : propertyMetadata.getFieldMetadataSet() ) { final FieldBridge fieldBridge = fieldMetadata.getFieldBridge(); final String fieldName = fieldMetadata.getName(); final FieldBridge oneWayConversionContext = conversionContext.oneWayConversionContext( fieldBridge ); // handle the default field creation via the bridge oneWayConversionContext.set( fieldName, currentFieldValue, document, typeMetadata.getFieldLuceneOptions( propertyMetadata, fieldMetadata, currentFieldValue, documentBoost ) ); // handle faceting fields if ( fieldMetadata.hasFacets() ) { faceting.enableFacetProcessing(); for ( FacetMetadata facetMetadata : fieldMetadata.getFacetMetadata() ) { if ( multiValued || isCollection ) { faceting.setMultiValued( facetMetadata.getFacetName() ); } addFacetDocValues( document, fieldMetadata, facetMetadata, currentFieldValue ); } } } // add the doc value fields required for sorting, but only if this property is not part of an embedded // to-many assoc, in which case sorting on these fields would not make sense if ( !multiValued ) { addSortFieldDocValues( document, propertyMetadata, documentBoost, currentFieldValue ); } } finally { conversionContext.popProperty(); } } } private void addFacetDocValues(Document document, DocumentFieldMetadata fieldMetadata, FacetMetadata facetMetadata, Object value) { // we don't add null values to the facet field if ( value == null ) { return; } List facetFields = new ArrayList<>(); Field field; if ( value instanceof Collection ) { for (String v : (Collection)value) { field = getFacetDocValues(document, fieldMetadata, facetMetadata, v); if (field != null) { facetFields.add(field); } } } else { field = getFacetDocValues(document, fieldMetadata, facetMetadata, value); if (field != null) { facetFields.add(field); } } for (Field facetField : facetFields) { document.add(facetField); } } private Field getFacetDocValues(Document document, DocumentFieldMetadata fieldMetadata, FacetMetadata facetMetadata, Object value) { // we don't add null values to the facet field if ( value == null ) { return null; } Field facetField; switch ( facetMetadata.getEncoding() ) { case STRING: { String stringValue = value.toString(); // we don't add empty strings to the facet field if ( stringValue.isEmpty() ) { return null; } facetField = new SortedSetDocValuesFacetField( facetMetadata.getFacetName(), stringValue ); break; } case LONG: { if ( value instanceof Number ) { facetField = new NumericDocValuesField( facetMetadata.getFacetName(), ( (Number) value ).longValue() ); } else if ( Date.class.isAssignableFrom( value.getClass() ) ) { Date date = (Date) value; FieldBridge fieldBridge = fieldMetadata.getFieldBridge(); // if we have a date and the actual value is not indexed as a numeric value we will run into // problems. Better fail early if ( !( fieldBridge instanceof NumericEncodingDateBridge ) ) { log.numericDateFacetForNonNumericField( facetMetadata.getFacetName(), fieldMetadata.getFieldName() ); } NumericEncodingDateBridge dateBridge = (NumericEncodingDateBridge) fieldBridge; long numericDateValue = DateTools.round( date.getTime(), dateBridge.getResolution() ); facetField = new NumericDocValuesField( facetMetadata.getFacetName(), numericDateValue ); } else if ( Calendar.class.isAssignableFrom( value.getClass() ) ) { Calendar calendar = (Calendar) value; facetField = new NumericDocValuesField( facetMetadata.getFacetName(), calendar.getTime().getTime() ); } else { throw new AssertionFailure( "Unexpected value type for faceting: " + value.getClass().getName() ); } break; } case DOUBLE: { if ( value instanceof Number ) { facetField = new DoubleDocValuesField( facetMetadata.getFacetName(), ( (Number) value ).doubleValue() ); } else { throw new AssertionFailure( "Unexpected value type for faceting: " + value.getClass().getName() ); } break; } case AUTO: { throw new AssertionFailure( "The facet type should have been resolved during bootstrapping" ); } default: { throw new AssertionFailure( "Unexpected facet encoding type '" + facetMetadata.getEncoding() + "' Has the enum been modified?" ); } } return facetField; } /** * Adds the doc field values to the document required to map the con value from the * underlying field will be obtained from the document (it has been written at this point already) and an equivalent * doc field value will be added. * For non-numeric fields, if the field value is supposed to be analyzed, the analysis will be performed by * this method, because Lucene does not analyze SortedDocValuesFields values automatically. */ private void addSortFieldDocValues(Document document, PropertyMetadata propertyMetadata, float documentBoost, Object propertyValue) { for ( SortableFieldMetadata sortField : propertyMetadata.getSortableFieldMetadata() ) { DocumentFieldMetadata fieldMetaData = propertyMetadata.getFieldMetadata( sortField.getFieldName() ); // field marked as sortable by custom bridge to allow sort field validation pass, but that bridge itself is // in charge of adding the required field if ( fieldMetaData == null ) { continue; } IndexableField field; // A non-stored, non-indexed field will not be added to the actual document; in that case retrieve // its value via a dummy document, adjusting the options to index the field if ( fieldMetaData.getIndex() == Index.NO && fieldMetaData.getStore() == Store.NO ) { FieldBridge fieldBridge = fieldMetaData.getFieldBridge(); LuceneOptionsImpl luceneOptions = new LuceneOptionsImpl( Index.NOT_ANALYZED, fieldMetaData.getTermVector(), fieldMetaData.getStore(), fieldMetaData.indexNullAs(), fieldMetaData.getBoost() * propertyMetadata.getDynamicBoostStrategy().defineBoost( propertyValue ), documentBoost ); Document dummy = new Document(); fieldBridge.set( "dummy", propertyValue, dummy, luceneOptions ); field = dummy.getField( "dummy" ); } else { field = document.getField( sortField.getFieldName() ); } if ( field != null ) { Number numericValue = field.numericValue(); if ( numericValue != null ) { if ( numericValue instanceof Double ) { document.add( new DoubleDocValuesField( sortField.getFieldName(), (double) numericValue ) ); } else if ( numericValue instanceof Float ) { document.add( new FloatDocValuesField( sortField.getFieldName(), (float) numericValue ) ); } else { document.add( new NumericDocValuesField( sortField.getFieldName(), numericValue.longValue() ) ); } } else { String value = field.stringValue(); Analyzer analyzer = fieldMetaData.getAnalyzer(); if ( fieldMetaData.getIndex().isAnalyzed() && analyzer != null ) { /* * Necessary because Lucene doesn't automatically analyze SortedDocValuesFields. * See https://hibernate.atlassian.net/browse/HSEARCH-2376 * * Analysis is skipped altogether when the analyzer is remote. It's up to the backend to handle it. */ value = InternalAnalyzerUtils.analyzeSortableValue( analyzer, sortField.getFieldName(), value ); } if ( value != null ) { document.add( new SortedDocValuesField( sortField.getFieldName(), new BytesRef( value ) ) ); } } } } } private void buildDocumentFieldForClassBridges(Document doc, TypeMetadata typeMetadata, ConversionContext conversionContext, float documentBoost, Object unproxiedInstance) { for ( DocumentFieldMetadata fieldMetadata : typeMetadata.getClassBridgeMetadata() ) { FieldBridge fieldBridge = fieldMetadata.getFieldBridge(); final String fieldName = fieldMetadata.getName(); final FieldBridge oneWayConversionContext = conversionContext.oneWayConversionContext( fieldBridge ); conversionContext.pushProperty( fieldName ); try { oneWayConversionContext.set( fieldName, unproxiedInstance, doc, typeMetadata.getClassLuceneOptions( fieldMetadata, documentBoost ) ); } finally { conversionContext.popProperty(); } } } /** * Check if a given value is present in an array. * * A array {@code null} contains all possible values. * Otherwise, a value {@code null} is always considered non present. * The last behavior is not currently used, it is merely an optimization. */ private boolean containsFieldName(String value, String[] array) { //null array means contains all possible values if ( array == null ) { return true; } //null value is meaningless if ( value == null ) { return false; } for ( String containedValue : array ) { if ( value.equals( containedValue ) ) { return true; } } return false; } private void processEmbeddedNullValue(Document doc, EmbeddedTypeMetadata embeddedTypeMetadata, ConversionContext conversionContext) { final String nullMarker = embeddedTypeMetadata.getEmbeddedNullToken(); if ( nullMarker != null ) { String fieldName = embeddedTypeMetadata.getEmbeddedNullFieldName(); FieldBridge fieldBridge = embeddedTypeMetadata.getEmbeddedNullFieldBridge(); final FieldBridge contextualizedBridge = conversionContext.oneWayConversionContext( fieldBridge ); conversionContext.pushProperty( fieldName ); try { contextualizedBridge.set( fieldName, null, doc, NULL_EMBEDDED_MARKER_OPTIONS ); } finally { conversionContext.popProperty(); } } } private Object unproxy(Object instance, InstanceInitializer objectInitializer) { if ( instance == null ) { return null; } return objectInitializer.unproxy( instance ); } /** * Allows a analyzer discriminator to override the analyzer used for any field in the Lucene document. * * @param doc The Lucene Document which shall be indexed. * @param typeMetadata The metadata for the entity we currently add to the document. * @param fieldToAnalyzerMap This map contains the actual override data. It is a map between document fields names and * analyzer definition names. This map will be added to the Work instance and processed at actual indexing time. * @param processedFieldNames A list of field names we have already processed. * @param unproxiedInstance The entity we currently "add" to the document. */ private void allowAnalyzerDiscriminatorOverride(Document doc, TypeMetadata typeMetadata, Map fieldToAnalyzerMap, Set processedFieldNames, Object unproxiedInstance) { Discriminator discriminator = typeMetadata.getDiscriminator(); if ( discriminator == null ) { return; } Object value = null; if ( typeMetadata.getDiscriminatorGetter() != null ) { value = ReflectionHelper.getMemberValue( unproxiedInstance, typeMetadata.getDiscriminatorGetter() ); } // now we give the discriminator the opportunity to specify a analyzer per field level for ( IndexableField field : doc.getFields() ) { if ( !processedFieldNames.contains( field.name() ) ) { String analyzerName = discriminator.getAnalyzerDefinitionName( value, unproxiedInstance, field.name() ); if ( analyzerName != null ) { fieldToAnalyzerMap.put( field.name(), analyzerName ); } processedFieldNames.add( field.name() ); } } } public String getIdentifierName() { return identifierName; } public boolean allowFieldSelectionInProjection() { return allowFieldSelectionInProjection; } /** * This method will be removed as Field caching is no longer implemented * (as it is no longer useful) * @return Always returns an empty Set. */ @Deprecated public Set getFieldCacheOption() { return Collections.emptySet(); } public TwoWayFieldBridge getIdBridge() { return (TwoWayFieldBridge) idPropertyMetadata.getFieldMetadata( idFieldName ).getFieldBridge(); } public String getIdKeywordName() { return idPropertyMetadata.getFieldMetadata( idFieldName ).getName(); } /** * Return the id used for indexing if possible * An IllegalStateException otherwise *

* If the id is provided, we can't extract it from the entity */ @Override public Serializable getId(Object entity) { if ( entity == null || idPropertyMetadata.getPropertyAccessorName() == null || idProvided ) { throw new IllegalStateException( "Cannot guess id from entity" ); } Object unproxiedEntity = getInstanceInitializer().unproxy( entity ); return (Serializable) ReflectionHelper.getMemberValue( unproxiedEntity, idPropertyMetadata.getPropertyAccessor() ); } public String objectToString(String fieldName, Object value, ConversionContext conversionContext) { if ( fieldName == null ) { throw new AssertionFailure( "Field name should not be null" ); } final DocumentFieldMetadata idFieldMetaData = idPropertyMetadata.getFieldMetadata( idFieldName ); final FieldBridge bridge = fieldName.equals( idFieldMetaData.getName() ) ? getIdBridge() : getBridge( getMetadata(), fieldName ); if ( bridge != null ) { return objectToString( fieldName, bridge, value, conversionContext ); } throw new SearchException( "Unable to find field " + fieldName + " in " + getBeanXClass() ); } public String objectToString(String fieldName, FieldBridge bridge, Object value, ConversionContext conversionContext) { if ( fieldName == null ) { throw new AssertionFailure( "Field name should not be null" ); } if ( bridge == null ) { throw new AssertionFailure( "Field bridge should not be null" ); } final Class bridgeClass = bridge.getClass(); if ( TwoWayFieldBridge.class.isAssignableFrom( bridgeClass ) ) { return objectToString( (TwoWayFieldBridge) bridge, fieldName, value, conversionContext ); } else if ( StringBridge.class.isAssignableFrom( bridgeClass ) ) { return objectToString( (StringBridge) bridge, fieldName, value, conversionContext ); } else { throw log.fieldBridgeNotTwoWay( bridgeClass, fieldName, getBeanXClass() ); } } private FieldBridge getNullBridge(EmbeddedTypeMetadata embeddedTypeMetadata, String fieldName) { if ( fieldName.equals( embeddedTypeMetadata.getEmbeddedNullFieldName() ) ) { return embeddedTypeMetadata.getEmbeddedNullFieldBridge(); } else { return null; } } public FieldBridge getBridge(String fieldName) { return getBridge( getMetadata(), fieldName ); } private FieldBridge getBridge(TypeMetadata typeMetadata, String fieldName) { // process base fields DocumentFieldMetadata documentFieldMetadata = typeMetadata.getDocumentFieldMetadataFor( fieldName ); if ( documentFieldMetadata != null && documentFieldMetadata.getFieldBridge() != null ) { return documentFieldMetadata.getFieldBridge(); } // process embedded fields FieldBridge fieldBridge; for ( EmbeddedTypeMetadata embeddedTypeMetadata : typeMetadata.getEmbeddedTypeMetadata() ) { fieldBridge = getBridge( embeddedTypeMetadata, fieldName ); if ( fieldBridge != null ) { return fieldBridge; } } // process null embedded fields for ( EmbeddedTypeMetadata embeddedTypeMetadata : typeMetadata.getEmbeddedTypeMetadata() ) { fieldBridge = getNullBridge( embeddedTypeMetadata, fieldName ); if ( fieldBridge != null ) { return fieldBridge; } } //process class bridges DocumentFieldMetadata fieldMetadata = typeMetadata.getFieldMetadataForClassBridgeField( fieldName ); fieldBridge = fieldMetadata == null ? null : fieldMetadata.getFieldBridge(); return fieldBridge; } /** * Checks whether all involved bridges allow to optimize document retrieval by using * {@code FieldSelector} (see HSEARCH-213). */ private void checkAllowFieldSelection() { allowFieldSelectionInProjection = true; if ( fieldBridgeProhibitsFieldSelectionInProjection( getIdBridge() ) ) { allowFieldSelectionInProjection = false; return; } for ( PropertyMetadata propertyMetadata : getMetadata().getAllPropertyMetadata() ) { for ( DocumentFieldMetadata documentFieldMetadata : propertyMetadata.getFieldMetadataSet() ) { FieldBridge bridge = documentFieldMetadata.getFieldBridge(); if ( fieldBridgeProhibitsFieldSelectionInProjection( bridge ) ) { allowFieldSelectionInProjection = false; return; } } } for ( DocumentFieldMetadata fieldMetadata : getMetadata().getClassBridgeMetadata() ) { FieldBridge bridge = fieldMetadata.getFieldBridge(); if ( fieldBridgeProhibitsFieldSelectionInProjection( bridge ) ) { allowFieldSelectionInProjection = false; return; } } } private boolean fieldBridgeProhibitsFieldSelectionInProjection(FieldBridge bridge) { return !( bridge instanceof TwoWayStringBridge || bridge instanceof TwoWayString2FieldBridgeAdaptor || bridge instanceof NumericFieldBridge || bridge instanceof NumericEncodingCalendarBridge || bridge instanceof NumericEncodingDateBridge || bridge instanceof StringEncodingDateBridge ); } /** * To be removed, see org.hibernate.search.engine.DocumentBuilderIndexedEntity.isIdMatchingJpaId() */ @Override public boolean requiresProvidedId() { return this.idProvided; } /** * FIXME remove the need for such a method, we should always be able to rely on Work.id, * but to respect @DocumentId which is being processed in the DocumentBuilder currently * finding out which id we need is tricky, and requires helpers method like this one. */ @Override public boolean isIdMatchingJpaId() { return ( !idProvided && getTypeMetadata().isJpaIdUsedAsDocumentId() ); } }