mirror of
https://github.com/Alfresco/SearchServices.git
synced 2025-09-17 14:21:20 +00:00
Merge branch 'feature/SEARCH-1949_ConfigOptionForPathQueries' into 'master'
Feature/search 1949 config option for path queries See merge request search_discovery/insightengine!343
This commit is contained in:
@@ -31,7 +31,7 @@ import static org.testng.Assert.assertTrue;
|
||||
* @author Alessandro Benedetti
|
||||
* @author Meenal Bhave
|
||||
*/
|
||||
public class CascadingTrackerIntegrationTest extends AbstractE2EFunctionalTest
|
||||
public class CascadingIntegrationTest extends AbstractE2EFunctionalTest
|
||||
{
|
||||
@Autowired
|
||||
protected DataContent dataContent;
|
@@ -523,7 +523,7 @@ public class AlfrescoSolrDataModel implements QueryConstants
|
||||
}
|
||||
catch (IOException e)
|
||||
{
|
||||
log.info("Failed to read shared properties fat " + propertiesFile.getAbsolutePath());
|
||||
log.info("Failed to read shared properties at " + propertiesFile.getAbsolutePath());
|
||||
}
|
||||
|
||||
return props;
|
||||
|
@@ -88,7 +88,7 @@ public interface InformationServer extends InformationServerCollectionProvider
|
||||
|
||||
void indexNode(Node node, boolean overwrite) throws IOException, AuthenticationException, JSONException;
|
||||
|
||||
void indexNodes(List<Node> nodes, boolean overwrite, boolean cascade) throws IOException, AuthenticationException, JSONException;
|
||||
void indexNodes(List<Node> nodes, boolean overwrite) throws IOException, AuthenticationException, JSONException;
|
||||
|
||||
void cascadeNodes(List<NodeMetaData> nodes, boolean overwrite) throws IOException, AuthenticationException, JSONException;
|
||||
|
||||
@@ -181,4 +181,11 @@ public interface InformationServer extends InformationServerCollectionProvider
|
||||
String getBaseUrl();
|
||||
|
||||
void flushContentStore() throws IOException;
|
||||
|
||||
/**
|
||||
* Check if cascade tracking is enabled.
|
||||
*
|
||||
* @return true if cascade tracking is enabled (note that this is the default behaviour if not specified in the properties file).
|
||||
*/
|
||||
boolean cascadeTrackingEnabled();
|
||||
}
|
||||
|
@@ -19,6 +19,7 @@
|
||||
package org.alfresco.solr;
|
||||
|
||||
import static java.util.Optional.ofNullable;
|
||||
|
||||
import static org.alfresco.repo.search.adaptor.lucene.QueryConstants.FIELD_ACLID;
|
||||
import static org.alfresco.repo.search.adaptor.lucene.QueryConstants.FIELD_ACLTXCOMMITTIME;
|
||||
import static org.alfresco.repo.search.adaptor.lucene.QueryConstants.FIELD_ACLTXID;
|
||||
@@ -76,17 +77,32 @@ import java.io.PrintWriter;
|
||||
import java.io.StringWriter;
|
||||
import java.nio.charset.StandardCharsets;
|
||||
import java.text.DecimalFormat;
|
||||
import java.util.*;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collection;
|
||||
import java.util.Collections;
|
||||
import java.util.EnumMap;
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
import java.util.Iterator;
|
||||
import java.util.LinkedHashMap;
|
||||
import java.util.LinkedHashSet;
|
||||
import java.util.LinkedList;
|
||||
import java.util.List;
|
||||
import java.util.Locale;
|
||||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
import java.util.Objects;
|
||||
import java.util.Properties;
|
||||
import java.util.Set;
|
||||
import java.util.concurrent.locks.ReentrantReadWriteLock;
|
||||
import java.util.regex.Matcher;
|
||||
import java.util.regex.Pattern;
|
||||
import java.util.zip.GZIPInputStream;
|
||||
|
||||
import com.carrotsearch.hppc.IntArrayList;
|
||||
|
||||
import com.carrotsearch.hppc.LongHashSet;
|
||||
import com.carrotsearch.hppc.cursors.LongCursor;
|
||||
|
||||
import org.alfresco.httpclient.AuthenticationException;
|
||||
import org.alfresco.model.ContentModel;
|
||||
import org.alfresco.opencmis.dictionary.CMISStrictDictionaryService;
|
||||
@@ -142,7 +158,19 @@ import org.apache.lucene.index.LeafReaderContext;
|
||||
import org.apache.lucene.index.NumericDocValues;
|
||||
import org.apache.lucene.index.ReaderUtil;
|
||||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.search.*;
|
||||
import org.apache.lucene.search.BooleanClause;
|
||||
import org.apache.lucene.search.BooleanQuery;
|
||||
import org.apache.lucene.search.Collector;
|
||||
import org.apache.lucene.search.LeafCollector;
|
||||
import org.apache.lucene.search.LegacyNumericRangeQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.ScoreDoc;
|
||||
import org.apache.lucene.search.Scorer;
|
||||
import org.apache.lucene.search.Sort;
|
||||
import org.apache.lucene.search.SortField;
|
||||
import org.apache.lucene.search.TermQuery;
|
||||
import org.apache.lucene.search.TopDocs;
|
||||
import org.apache.lucene.search.TopFieldCollector;
|
||||
import org.apache.lucene.util.BytesRefBuilder;
|
||||
import org.apache.solr.common.SolrDocument;
|
||||
import org.apache.solr.common.SolrDocumentList;
|
||||
@@ -167,8 +195,6 @@ import org.apache.solr.search.DelegatingCollector;
|
||||
import org.apache.solr.search.DocIterator;
|
||||
import org.apache.solr.search.DocList;
|
||||
import org.apache.solr.search.DocSet;
|
||||
import org.apache.solr.search.QueryCommand;
|
||||
import org.apache.solr.search.QueryResult;
|
||||
import org.apache.solr.search.QueryWrapperFilter;
|
||||
import org.apache.solr.search.SolrIndexSearcher;
|
||||
import org.apache.solr.update.AddUpdateCommand;
|
||||
@@ -205,10 +231,7 @@ public class SolrInformationServer implements InformationServer
|
||||
|
||||
public static final String AND = " AND ";
|
||||
public static final String OR = " OR ";
|
||||
//public static final String REQUEST_HANDLER_ALFRESCO_FULL_TEXT_SEARCH = "/afts";
|
||||
private static final String REQUEST_HANDLER_NATIVE = "/native";
|
||||
//public static final String REQUEST_HANDLER_ALFRESCO = "/alfresco";
|
||||
//public static final String REQUEST_HANDLER_SELECT = "/select";
|
||||
static final String REQUEST_HANDLER_GET = "/get";
|
||||
private static final String RESPONSE_DEFAULT_IDS = "response";
|
||||
static final String RESPONSE_DEFAULT_ID = "doc";
|
||||
@@ -232,6 +255,8 @@ public class SolrInformationServer implements InformationServer
|
||||
*/
|
||||
private static final int BATCH_FACET_TXS = 4096;
|
||||
private static final String FINGERPRINT_FIELD = "MINHASH";
|
||||
/** Shared property to determine if the cascade tracking is enabled. */
|
||||
public static final String CASCADE_TRACKER_ENABLED = "alfresco.cascade.tracker.enabled";
|
||||
|
||||
private final AlfrescoCoreAdminHandler adminHandler;
|
||||
private final SolrCore core;
|
||||
@@ -441,7 +466,7 @@ public class SolrInformationServer implements InformationServer
|
||||
dataModel = AlfrescoSolrDataModel.getInstance();
|
||||
|
||||
contentStreamLimit = Integer.parseInt(p.getProperty("alfresco.contentStreamLimit", "10000000"));
|
||||
|
||||
|
||||
// build base URL - host and port have to come from configuration.
|
||||
props = AlfrescoSolrDataModel.getCommonConfig();
|
||||
hostName = ConfigUtil.locateProperty(SOLR_HOST, props.getProperty(SOLR_HOST));
|
||||
@@ -456,6 +481,14 @@ public class SolrInformationServer implements InformationServer
|
||||
return this.adminHandler;
|
||||
}
|
||||
|
||||
/** {@inheritDoc} */
|
||||
@Override
|
||||
public boolean cascadeTrackingEnabled()
|
||||
{
|
||||
String cascadeTrackerEnabledProp = ofNullable((String) props.get(CASCADE_TRACKER_ENABLED)).orElse("true");
|
||||
return Boolean.valueOf(cascadeTrackerEnabledProp);
|
||||
}
|
||||
|
||||
@Override
|
||||
public synchronized void initSkippingDescendantDocs()
|
||||
{
|
||||
@@ -507,7 +540,7 @@ public class SolrInformationServer implements InformationServer
|
||||
report.add("Node count with FTSStatus New", newCount);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public void afterInitModels()
|
||||
{
|
||||
@@ -520,7 +553,7 @@ public class SolrInformationServer implements InformationServer
|
||||
String query = FIELD_ACLID + ":" + aclid + AND + FIELD_DOC_TYPE + ":" + DOC_TYPE_ACL;
|
||||
long count = this.getDocListSize(query);
|
||||
aclReport.setIndexedAclDocCount(count);
|
||||
|
||||
|
||||
// TODO Could add INACLTXID later, but would need acl change set id.
|
||||
return aclReport;
|
||||
}
|
||||
@@ -778,7 +811,7 @@ public class SolrInformationServer implements InformationServer
|
||||
long count = this.getDocListSize(query);
|
||||
nodeReport.setIndexedNodeDocCount(count);
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public void commit() throws IOException
|
||||
{
|
||||
@@ -891,26 +924,26 @@ public class SolrInformationServer implements InformationServer
|
||||
|
||||
return searcherOpened;
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public void deleteByAclChangeSetId(Long aclChangeSetId) throws IOException
|
||||
{
|
||||
deleteById(FIELD_INACLTXID, aclChangeSetId);
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public void deleteByAclId(Long aclId) throws IOException
|
||||
{
|
||||
isIdIndexCache.clear();
|
||||
deleteById(FIELD_ACLID, aclId);
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public void deleteByNodeId(Long nodeId) throws IOException
|
||||
{
|
||||
deleteById(FIELD_DBID, nodeId);
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public void deleteByTransactionId(Long transactionId) throws IOException
|
||||
{
|
||||
@@ -923,7 +956,7 @@ public class SolrInformationServer implements InformationServer
|
||||
{
|
||||
return this.dataModel.getAlfrescoModels();
|
||||
}
|
||||
|
||||
|
||||
@SuppressWarnings({ "unchecked", "rawtypes" })
|
||||
@Override
|
||||
public Iterable<Entry<String, Object>> getCoreStats() throws IOException
|
||||
@@ -1047,7 +1080,7 @@ public class SolrInformationServer implements InformationServer
|
||||
|
||||
return coreSummary;
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public DictionaryComponent getDictionaryService(String alternativeDictionary)
|
||||
{
|
||||
@@ -1371,7 +1404,10 @@ public class SolrInformationServer implements InformationServer
|
||||
public void dirtyTransaction(long txnId)
|
||||
{
|
||||
this.cleanContentCache.remove(txnId);
|
||||
this.cleanCascadeCache.remove(txnId);
|
||||
if (cascadeTrackingEnabled())
|
||||
{
|
||||
this.cleanCascadeCache.remove(txnId);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
@@ -1460,9 +1496,9 @@ public class SolrInformationServer implements InformationServer
|
||||
long start = System.nanoTime();
|
||||
|
||||
if ((node.getStatus() == SolrApiNodeStatus.DELETED)
|
||||
|| (node.getStatus() == SolrApiNodeStatus.NON_SHARD_DELETED)
|
||||
|| (node.getStatus() == SolrApiNodeStatus.NON_SHARD_UPDATED)
|
||||
|| (node.getStatus() == SolrApiNodeStatus.UNKNOWN))
|
||||
|| (node.getStatus() == SolrApiNodeStatus.UNKNOWN)
|
||||
|| cascadeTrackingEnabled() && ((node.getStatus() == SolrApiNodeStatus.NON_SHARD_DELETED)
|
||||
|| (node.getStatus() == SolrApiNodeStatus.NON_SHARD_UPDATED)))
|
||||
{
|
||||
// fix up any secondary paths
|
||||
NodeMetaDataParameters nmdp = new NodeMetaDataParameters();
|
||||
@@ -1470,8 +1506,8 @@ public class SolrInformationServer implements InformationServer
|
||||
nmdp.setToNodeId(node.getId());
|
||||
List<NodeMetaData> nodeMetaDatas;
|
||||
if ((node.getStatus() == SolrApiNodeStatus.DELETED)
|
||||
|| (node.getStatus() == SolrApiNodeStatus.NON_SHARD_DELETED)
|
||||
|| (node.getStatus() == SolrApiNodeStatus.NON_SHARD_UPDATED))
|
||||
|| cascadeTrackingEnabled() && ((node.getStatus() == SolrApiNodeStatus.NON_SHARD_DELETED)
|
||||
|| (node.getStatus() == SolrApiNodeStatus.NON_SHARD_UPDATED)))
|
||||
{
|
||||
// Fake the empty node metadata for this parent deleted node
|
||||
NodeMetaData nodeMetaData = createDeletedNodeMetaData(node);
|
||||
@@ -1481,7 +1517,7 @@ public class SolrInformationServer implements InformationServer
|
||||
{
|
||||
nodeMetaDatas = repositoryClient.getNodesMetaData(nmdp, Integer.MAX_VALUE);
|
||||
}
|
||||
|
||||
|
||||
NodeMetaData nodeMetaData;
|
||||
if (!nodeMetaDatas.isEmpty())
|
||||
{
|
||||
@@ -1499,7 +1535,7 @@ public class SolrInformationServer implements InformationServer
|
||||
finally
|
||||
{
|
||||
unlock(nodeMetaData.getId());
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
// else, the node has moved on to a later transaction, and it will be indexed later
|
||||
@@ -1508,10 +1544,9 @@ public class SolrInformationServer implements InformationServer
|
||||
deleteNode(processor, request, node);
|
||||
}
|
||||
|
||||
|
||||
if ((node.getStatus() == SolrApiNodeStatus.UPDATED)
|
||||
|| (node.getStatus() == SolrApiNodeStatus.UNKNOWN)
|
||||
|| (node.getStatus() == SolrApiNodeStatus.NON_SHARD_UPDATED))
|
||||
if (node.getStatus() == SolrApiNodeStatus.UPDATED
|
||||
|| node.getStatus() == SolrApiNodeStatus.UNKNOWN
|
||||
|| (cascadeTrackingEnabled() && node.getStatus() == SolrApiNodeStatus.NON_SHARD_UPDATED))
|
||||
{
|
||||
long nodeId = node.getId();
|
||||
|
||||
@@ -1785,7 +1820,7 @@ public class SolrInformationServer implements InformationServer
|
||||
|
||||
|
||||
@Override
|
||||
public void indexNodes(List<Node> nodes, boolean overwrite, boolean cascade) throws IOException, JSONException
|
||||
public void indexNodes(List<Node> nodes, boolean overwrite) throws IOException, JSONException
|
||||
{
|
||||
UpdateRequestProcessor processor = null;
|
||||
try (SolrQueryRequest request = newSolrQueryRequest())
|
||||
@@ -1796,8 +1831,13 @@ public class SolrInformationServer implements InformationServer
|
||||
EnumMap<SolrApiNodeStatus, List<Long>> nodeStatusToNodeIds = new EnumMap<>(SolrApiNodeStatus.class);
|
||||
categorizeNodes(nodes, nodeIdsToNodes, nodeStatusToNodeIds);
|
||||
List<Long> deletedNodeIds = mapNullToEmptyList(nodeStatusToNodeIds.get(SolrApiNodeStatus.DELETED));
|
||||
List<Long> shardDeletedNodeIds = mapNullToEmptyList(nodeStatusToNodeIds.get(SolrApiNodeStatus.NON_SHARD_DELETED));
|
||||
List<Long> shardUpdatedNodeIds = mapNullToEmptyList(nodeStatusToNodeIds.get(SolrApiNodeStatus.NON_SHARD_UPDATED));
|
||||
List<Long> shardDeletedNodeIds = Collections.emptyList();
|
||||
List<Long> shardUpdatedNodeIds = Collections.emptyList();
|
||||
if (cascadeTrackingEnabled())
|
||||
{
|
||||
shardDeletedNodeIds = mapNullToEmptyList(nodeStatusToNodeIds.get(SolrApiNodeStatus.NON_SHARD_DELETED));
|
||||
shardUpdatedNodeIds = mapNullToEmptyList(nodeStatusToNodeIds.get(SolrApiNodeStatus.NON_SHARD_UPDATED));
|
||||
}
|
||||
List<Long> unknownNodeIds = mapNullToEmptyList(nodeStatusToNodeIds.get(SolrApiNodeStatus.UNKNOWN));
|
||||
List<Long> updatedNodeIds = mapNullToEmptyList(nodeStatusToNodeIds.get(SolrApiNodeStatus.UPDATED));
|
||||
|
||||
@@ -1881,7 +1921,7 @@ public class SolrInformationServer implements InformationServer
|
||||
continue;
|
||||
}
|
||||
|
||||
if (nodeIdsToNodes.get(nodeMetaData.getId()).getStatus() == SolrApiNodeStatus.NON_SHARD_UPDATED)
|
||||
if (cascadeTrackingEnabled() && nodeIdsToNodes.get(nodeMetaData.getId()).getStatus() == SolrApiNodeStatus.NON_SHARD_UPDATED)
|
||||
{
|
||||
if (nodeMetaData.getProperties().get(ContentModel.PROP_CASCADE_TX) != null)
|
||||
{
|
||||
@@ -1982,7 +2022,7 @@ public class SolrInformationServer implements InformationServer
|
||||
{
|
||||
StringPropertyValue latProp = ((StringPropertyValue)nodeMetaData.getProperties().get(ContentModel.PROP_LATITUDE));
|
||||
StringPropertyValue lonProp = ((StringPropertyValue)nodeMetaData.getProperties().get(ContentModel.PROP_LONGITUDE));
|
||||
|
||||
|
||||
if((latProp != null) && (lonProp != null))
|
||||
{
|
||||
String lat = latProp.getValue();
|
||||
@@ -2020,14 +2060,14 @@ public class SolrInformationServer implements InformationServer
|
||||
}
|
||||
doc.addField(FIELD_ISNODE, "T");
|
||||
// FIELD_FTSSTATUS is set when adding content properties to indicate whether or not the cache is clean.
|
||||
|
||||
|
||||
doc.addField(FIELD_TENANT, AlfrescoSolrDataModel.getTenantId(nodeMetaData.getTenantDomain()));
|
||||
|
||||
updatePathRelatedFields(nodeMetaData, doc);
|
||||
updateNamePathRelatedFields(nodeMetaData, doc);
|
||||
updateAncestorRelatedFields(nodeMetaData, doc);
|
||||
doc.addField(FIELD_PARENT_ASSOC_CRC, nodeMetaData.getParentAssocsCrc());
|
||||
|
||||
|
||||
if (nodeMetaData.getOwner() != null)
|
||||
{
|
||||
doc.addField(FIELD_OWNER, nodeMetaData.getOwner());
|
||||
@@ -2113,7 +2153,7 @@ public class SolrInformationServer implements InformationServer
|
||||
}
|
||||
}
|
||||
|
||||
static void addPropertiesToDoc(Map<QName, PropertyValue> properties, boolean isContentIndexedForNode,
|
||||
static void addPropertiesToDoc(Map<QName, PropertyValue> properties, boolean isContentIndexedForNode,
|
||||
SolrInputDocument newDoc, SolrInputDocument cachedDoc, boolean transformContentFlag)
|
||||
{
|
||||
for (Entry<QName, PropertyValue> property : properties.entrySet())
|
||||
@@ -2121,7 +2161,7 @@ public class SolrInformationServer implements InformationServer
|
||||
QName propertyQName = property.getKey();
|
||||
newDoc.addField(FIELD_PROPERTIES, propertyQName.toString());
|
||||
newDoc.addField(FIELD_PROPERTIES, propertyQName.getPrefixString());
|
||||
|
||||
|
||||
PropertyValue value = property.getValue();
|
||||
if(value != null)
|
||||
{
|
||||
@@ -2195,22 +2235,22 @@ public class SolrInformationServer implements InformationServer
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
private void deleteErrorNode(UpdateRequestProcessor processor, SolrQueryRequest request, Node node) throws IOException
|
||||
{
|
||||
|
||||
|
||||
String errorDocId = PREFIX_ERROR + node.getId();
|
||||
|
||||
|
||||
// Try finding the node before performing removal operation
|
||||
DocSet docSet = request.getSearcher().getDocSet(new TermQuery(new Term(FIELD_SOLR4_ID, errorDocId)));
|
||||
|
||||
|
||||
if (docSet.size() > 0)
|
||||
{
|
||||
DeleteUpdateCommand delErrorDocCmd = new DeleteUpdateCommand(request);
|
||||
delErrorDocCmd.setId(errorDocId);
|
||||
processor.processDelete(delErrorDocCmd);
|
||||
}
|
||||
|
||||
|
||||
}
|
||||
|
||||
private void deleteNode(UpdateRequestProcessor processor, SolrQueryRequest request, Node node) throws IOException
|
||||
@@ -2221,20 +2261,20 @@ public class SolrInformationServer implements InformationServer
|
||||
// MNT-13767 fix, remove by node DBID.
|
||||
deleteNode(processor, request, node.getId());
|
||||
}
|
||||
|
||||
|
||||
private void deleteNode(UpdateRequestProcessor processor, SolrQueryRequest request, long dbid) throws IOException
|
||||
{
|
||||
|
||||
|
||||
// Try finding the node before performing removal operation
|
||||
DocSet docSet = request.getSearcher().getDocSet(LongPoint.newExactQuery(FIELD_DBID, dbid));
|
||||
|
||||
|
||||
if (docSet.size() > 0)
|
||||
{
|
||||
DeleteUpdateCommand delDocCmd = new DeleteUpdateCommand(request);
|
||||
delDocCmd.setQuery(FIELD_DBID + ":" + dbid);
|
||||
processor.processDelete(delDocCmd);
|
||||
}
|
||||
|
||||
|
||||
}
|
||||
|
||||
private boolean isContentIndexedForNode(Map<QName, PropertyValue> properties)
|
||||
@@ -2296,7 +2336,7 @@ public class SolrInformationServer implements InformationServer
|
||||
return fieldName;
|
||||
}
|
||||
|
||||
private void addContentPropertyMetadata(SolrInputDocument doc, QName propertyQName,
|
||||
private void addContentPropertyMetadata(SolrInputDocument doc, QName propertyQName,
|
||||
AlfrescoSolrDataModel.ContentFieldType type, GetTextContentResponse textContentResponse)
|
||||
{
|
||||
IndexedField indexedField = AlfrescoSolrDataModel.getInstance().getIndexedFieldForContentPropertyMetadata(
|
||||
@@ -2322,7 +2362,7 @@ public class SolrInformationServer implements InformationServer
|
||||
}
|
||||
}
|
||||
|
||||
private static void addContentPropertyMetadata(SolrInputDocument doc, QName propertyQName,
|
||||
private static void addContentPropertyMetadata(SolrInputDocument doc, QName propertyQName,
|
||||
ContentPropertyValue contentPropertyValue, AlfrescoSolrDataModel.ContentFieldType type)
|
||||
{
|
||||
IndexedField indexedField = AlfrescoSolrDataModel.getInstance().getIndexedFieldForContentPropertyMetadata(
|
||||
@@ -2352,8 +2392,8 @@ public class SolrInformationServer implements InformationServer
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private static void addContentPropertyToDocUsingCache(SolrInputDocument newDoc, SolrInputDocument cachedDoc,
|
||||
|
||||
private static void addContentPropertyToDocUsingCache(SolrInputDocument newDoc, SolrInputDocument cachedDoc,
|
||||
QName propertyQName, ContentPropertyValue contentPropertyValue, boolean transformContentFlag)
|
||||
{
|
||||
addContentPropertyMetadata(newDoc, propertyQName, contentPropertyValue, AlfrescoSolrDataModel.ContentFieldType.DOCID);
|
||||
@@ -2361,14 +2401,14 @@ public class SolrInformationServer implements InformationServer
|
||||
addContentPropertyMetadata(newDoc, propertyQName, contentPropertyValue, AlfrescoSolrDataModel.ContentFieldType.LOCALE);
|
||||
addContentPropertyMetadata(newDoc, propertyQName, contentPropertyValue, AlfrescoSolrDataModel.ContentFieldType.MIMETYPE);
|
||||
addContentPropertyMetadata(newDoc, propertyQName, contentPropertyValue, AlfrescoSolrDataModel.ContentFieldType.ENCODING);
|
||||
|
||||
|
||||
if (!transformContentFlag)
|
||||
{
|
||||
// Marks it as Clean so we do not get the actual content
|
||||
markFTSStatus(newDoc, FTSStatus.Clean);
|
||||
return;
|
||||
}
|
||||
|
||||
|
||||
if (cachedDoc != null)
|
||||
{
|
||||
ofNullable(cachedDoc.getField("MINHASH"))
|
||||
@@ -2386,7 +2426,7 @@ public class SolrInformationServer implements InformationServer
|
||||
addFieldIfNotSet(newDoc, field);
|
||||
}
|
||||
|
||||
String transformationStatusFieldName = getSolrFieldNameForContentPropertyMetadata(propertyQName,
|
||||
String transformationStatusFieldName = getSolrFieldNameForContentPropertyMetadata(propertyQName,
|
||||
AlfrescoSolrDataModel.ContentFieldType.TRANSFORMATION_STATUS);
|
||||
if (transformationStatusFieldName != null){
|
||||
newDoc.addField(transformationStatusFieldName, cachedDoc.getFieldValue(transformationStatusFieldName));
|
||||
@@ -2405,19 +2445,19 @@ public class SolrInformationServer implements InformationServer
|
||||
// Gets the new content docid and compares to that of the cachedDoc to mark the content as clean/dirty
|
||||
String fldName = getSolrFieldNameForContentPropertyMetadata(propertyQName,
|
||||
AlfrescoSolrDataModel.ContentFieldType.DOCID);
|
||||
|
||||
|
||||
if(newDoc.getFieldValue(FIELD_FTSSTATUS) == null)
|
||||
{
|
||||
newDoc.addField(FIELD_FTSSTATUS, cachedDoc.getFieldValue(FIELD_FTSSTATUS));
|
||||
}
|
||||
|
||||
|
||||
if(cachedDoc.getFieldValue(fldName) != null)
|
||||
{
|
||||
long cachedDocContentDocid = Long.parseLong(String.valueOf(cachedDoc.getFieldValue(fldName)));
|
||||
long currentContentDocid = contentPropertyValue.getId();
|
||||
// If we have used out of date content we mark it as dirty
|
||||
// Otherwise we leave it alone - it could already be marked as dirty/New and require an update
|
||||
|
||||
|
||||
if (cachedDocContentDocid != currentContentDocid)
|
||||
{
|
||||
// The cached content is out of date
|
||||
@@ -2429,7 +2469,7 @@ public class SolrInformationServer implements InformationServer
|
||||
markFTSStatus(newDoc, FTSStatus.Dirty);
|
||||
}
|
||||
}
|
||||
else
|
||||
else
|
||||
{
|
||||
// There is not a SolrInputDocument in the solrContentStore, so no content is added now to the new solr doc
|
||||
markFTSStatus(newDoc, FTSStatus.New);
|
||||
@@ -2468,7 +2508,7 @@ public class SolrInformationServer implements InformationServer
|
||||
doc.removeField(FIELD_FTSSTATUS);
|
||||
doc.addField(FIELD_FTSSTATUS, status.toString());
|
||||
}
|
||||
|
||||
|
||||
private void addContentToDoc(SolrInputDocument doc, long dbId) throws AuthenticationException, IOException
|
||||
{
|
||||
Collection<String> fieldNames = doc.deepCopy().getFieldNames();
|
||||
@@ -2484,22 +2524,22 @@ public class SolrInformationServer implements InformationServer
|
||||
// Could update multi content but it is broken ....
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
private void addContentPropertyToDocUsingAlfrescoRepository(SolrInputDocument doc,
|
||||
QName propertyQName, long dbId, String locale) throws AuthenticationException, IOException
|
||||
{
|
||||
long start = System.nanoTime();
|
||||
|
||||
|
||||
// Expensive call to be done with ContentTracker
|
||||
GetTextContentResponse response = repositoryClient.getTextContent(dbId, propertyQName, null);
|
||||
|
||||
|
||||
addContentPropertyMetadata(doc, propertyQName, AlfrescoSolrDataModel.ContentFieldType.TRANSFORMATION_STATUS,
|
||||
response);
|
||||
addContentPropertyMetadata(doc, propertyQName, AlfrescoSolrDataModel.ContentFieldType.TRANSFORMATION_EXCEPTION,
|
||||
response);
|
||||
addContentPropertyMetadata(doc, propertyQName, AlfrescoSolrDataModel.ContentFieldType.TRANSFORMATION_TIME,
|
||||
response);
|
||||
|
||||
|
||||
InputStream ris = response.getContent();
|
||||
if (Objects.equals(response.getContentEncoding(), "gzip"))
|
||||
{
|
||||
@@ -2545,7 +2585,7 @@ public class SolrInformationServer implements InformationServer
|
||||
|
||||
long end = System.nanoTime();
|
||||
this.getTrackerStats().addDocTransformationTime(end - start);
|
||||
|
||||
|
||||
StringBuilder builder = new StringBuilder(textContent.length() + 16);
|
||||
builder.append("\u0000").append(locale).append("\u0000");
|
||||
builder.append(textContent);
|
||||
@@ -2567,7 +2607,7 @@ public class SolrInformationServer implements InformationServer
|
||||
}
|
||||
|
||||
private static void addMLTextPropertyToDoc(SolrInputDocument doc, FieldInstance field, MLTextPropertyValue mlTextPropertyValue)
|
||||
{
|
||||
{
|
||||
if(field.isLocalised())
|
||||
{
|
||||
StringBuilder sort = new StringBuilder(128);
|
||||
@@ -2575,20 +2615,20 @@ public class SolrInformationServer implements InformationServer
|
||||
{
|
||||
final String propValue = mlTextPropertyValue.getValue(locale);
|
||||
LOGGER.debug("ML {} in {} of {}", field.getField(), locale, propValue);
|
||||
|
||||
|
||||
if((locale == null) || (propValue == null))
|
||||
{
|
||||
continue;
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
StringBuilder builder = new StringBuilder(propValue.length() + 16);
|
||||
builder.append("\u0000").append(locale.toString()).append("\u0000").append(propValue);
|
||||
|
||||
|
||||
if(!field.isSort())
|
||||
{
|
||||
doc.addField(field.getField(), builder.toString());
|
||||
}
|
||||
|
||||
|
||||
if (sort.length() > 0)
|
||||
{
|
||||
sort.append("\u0000");
|
||||
@@ -2629,7 +2669,10 @@ public class SolrInformationServer implements InformationServer
|
||||
input.addField(FIELD_INTXID, txn.getId());
|
||||
input.addField(FIELD_TXCOMMITTIME, txn.getCommitTimeMs());
|
||||
input.addField(FIELD_DOC_TYPE, DOC_TYPE_TX);
|
||||
input.addField(FIELD_CASCADE_FLAG, 0);
|
||||
if (cascadeTrackingEnabled())
|
||||
{
|
||||
input.addField(FIELD_CASCADE_FLAG, 0);
|
||||
}
|
||||
cmd.solrDoc = input;
|
||||
processor.processAdd(cmd);
|
||||
}
|
||||
@@ -2670,8 +2713,11 @@ public class SolrInformationServer implements InformationServer
|
||||
input.addField(FIELD_S_TXID, info.getId());
|
||||
input.addField(FIELD_S_TXCOMMITTIME, info.getCommitTimeMs());
|
||||
|
||||
//Set the cascade flag to 1. This means cascading updates have not been done yet.
|
||||
input.addField(FIELD_CASCADE_FLAG, 1);
|
||||
if (cascadeTrackingEnabled())
|
||||
{
|
||||
//Set the cascade flag to 1. This means cascading updates have not been done yet.
|
||||
input.addField(FIELD_CASCADE_FLAG, 1);
|
||||
}
|
||||
|
||||
cmd.solrDoc = input;
|
||||
processor.processAdd(cmd);
|
||||
@@ -2898,7 +2944,7 @@ public class SolrInformationServer implements InformationServer
|
||||
{
|
||||
activeTrackerThreadsLock.writeLock().unlock();
|
||||
}
|
||||
|
||||
|
||||
}
|
||||
|
||||
@Override
|
||||
@@ -2925,7 +2971,7 @@ public class SolrInformationServer implements InformationServer
|
||||
SolrIndexSearcher solrIndexSearcher = refCounted.get();
|
||||
|
||||
NumericDocValues dbidDocValues = solrIndexSearcher.getSlowAtomicReader().getNumericDocValues(QueryConstants.FIELD_DBID);
|
||||
|
||||
|
||||
List<Node> batch = new ArrayList<>(200);
|
||||
DocList docList = cloud.getDocList(nativeRequestHandler, request, query.startsWith("{") ? query : "{!afts}"+query);
|
||||
for (DocIterator it = docList.iterator(); it.hasNext(); /**/)
|
||||
@@ -2940,16 +2986,16 @@ public class SolrInformationServer implements InformationServer
|
||||
node.setTxnId(Long.MAX_VALUE);
|
||||
|
||||
batch.add(node);
|
||||
|
||||
|
||||
if(batch.size() >= 200)
|
||||
{
|
||||
indexNodes(batch, true, true);
|
||||
indexNodes(batch, true);
|
||||
batch.clear();
|
||||
}
|
||||
}
|
||||
if(batch.size() > 0)
|
||||
{
|
||||
indexNodes(batch, true, true);
|
||||
indexNodes(batch, true);
|
||||
batch.clear();
|
||||
}
|
||||
}
|
||||
@@ -3438,7 +3484,7 @@ public class SolrInformationServer implements InformationServer
|
||||
SolrQueryRequest request, UpdateRequestProcessor processor, LinkedHashSet<Long> stack)
|
||||
throws AuthenticationException, IOException, JSONException
|
||||
{
|
||||
|
||||
|
||||
// skipDescendantDocsForSpecificAspects is initialised on a synchronised method, so access must be also synchronised
|
||||
synchronized (this)
|
||||
{
|
||||
|
@@ -21,6 +21,15 @@ package org.alfresco.solr.lifecycle;
|
||||
import static java.util.Arrays.asList;
|
||||
import static java.util.Optional.ofNullable;
|
||||
|
||||
import static org.alfresco.solr.SolrInformationServer.CASCADE_TRACKER_ENABLED;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collection;
|
||||
import java.util.List;
|
||||
import java.util.Properties;
|
||||
import java.util.function.Function;
|
||||
import java.util.function.Predicate;
|
||||
|
||||
import org.alfresco.opencmis.dictionary.CMISStrictDictionaryService;
|
||||
import org.alfresco.solr.AlfrescoCoreAdminHandler;
|
||||
import org.alfresco.solr.AlfrescoSolrDataModel;
|
||||
@@ -54,13 +63,6 @@ import org.apache.solr.search.SolrIndexSearcher;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collection;
|
||||
import java.util.List;
|
||||
import java.util.Properties;
|
||||
import java.util.function.Function;
|
||||
import java.util.function.Predicate;
|
||||
|
||||
/**
|
||||
* Listeners for *FIRST SEARCHER* events in order to prepare and register the SolrContentStore and the Tracking Subsystem.
|
||||
*
|
||||
@@ -257,19 +259,27 @@ public class SolrCoreLoadListener extends AbstractSolrEventListener
|
||||
trackerRegistry,
|
||||
scheduler);
|
||||
|
||||
CascadeTracker cascadeTracker =
|
||||
registerAndSchedule(
|
||||
new CascadeTracker(props, repositoryClient, core.getName(), srv),
|
||||
core,
|
||||
props,
|
||||
trackerRegistry,
|
||||
scheduler);
|
||||
List<Tracker> trackers = new ArrayList<>();
|
||||
|
||||
String cascadeTrackerEnabledProp = ofNullable((String) props.get(CASCADE_TRACKER_ENABLED)).orElse("true");
|
||||
if (Boolean.valueOf(cascadeTrackerEnabledProp))
|
||||
{
|
||||
CascadeTracker cascadeTracker =
|
||||
registerAndSchedule(
|
||||
new CascadeTracker(props, repositoryClient, core.getName(), srv),
|
||||
core,
|
||||
props,
|
||||
trackerRegistry,
|
||||
scheduler);
|
||||
trackers.add(cascadeTracker);
|
||||
}
|
||||
|
||||
//The CommitTracker will acquire these locks in order
|
||||
//The ContentTracker will likely have the longest runs so put it first to ensure the MetadataTracker is not paused while
|
||||
//waiting for the ContentTracker to release it's lock.
|
||||
//The aclTracker will likely have the shortest runs so put it last.
|
||||
return asList(cascadeTracker, contentTracker, metadataTracker, aclTracker);
|
||||
trackers.addAll(asList(contentTracker, metadataTracker, aclTracker));
|
||||
return trackers;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@@ -19,7 +19,11 @@
|
||||
|
||||
package org.alfresco.solr.tracker;
|
||||
|
||||
import static java.util.Optional.empty;
|
||||
import static java.util.Optional.ofNullable;
|
||||
|
||||
import java.util.List;
|
||||
import java.util.Optional;
|
||||
import java.util.Properties;
|
||||
import java.util.concurrent.atomic.AtomicInteger;
|
||||
|
||||
@@ -41,7 +45,8 @@ public class CommitTracker extends AbstractTracker
|
||||
private MetadataTracker metadataTracker;
|
||||
private AclTracker aclTracker;
|
||||
private ContentTracker contentTracker;
|
||||
private CascadeTracker cascadeTracker;
|
||||
/** The cascade tracker. Note that this may be empty if cascade tracking is disabled. */
|
||||
private Optional<CascadeTracker> cascadeTracker = empty();
|
||||
private AtomicInteger rollbackCount = new AtomicInteger(0);
|
||||
|
||||
protected final static Logger log = LoggerFactory.getLogger(CommitTracker.class);
|
||||
@@ -71,7 +76,7 @@ public class CommitTracker extends AbstractTracker
|
||||
} else if(tracker instanceof ContentTracker) {
|
||||
this.contentTracker = (ContentTracker)tracker;
|
||||
} else if(tracker instanceof CascadeTracker) {
|
||||
this.cascadeTracker = (CascadeTracker)tracker;
|
||||
this.cascadeTracker = ofNullable((CascadeTracker) tracker);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -182,8 +187,11 @@ public class CommitTracker extends AbstractTracker
|
||||
contentTracker.getWriteLock().acquire();
|
||||
assert(contentTracker.getWriteLock().availablePermits() == 0);
|
||||
|
||||
cascadeTracker.getWriteLock().acquire();
|
||||
assert(cascadeTracker.getWriteLock().availablePermits() == 0);
|
||||
if (cascadeTracker.isPresent())
|
||||
{
|
||||
cascadeTracker.get().getWriteLock().acquire();
|
||||
assert (cascadeTracker.get().getWriteLock().availablePermits() == 0);
|
||||
}
|
||||
|
||||
infoSrv.rollback();
|
||||
}
|
||||
@@ -206,12 +214,12 @@ public class CommitTracker extends AbstractTracker
|
||||
contentTracker.invalidateState();
|
||||
|
||||
//Reset cascadeTracker
|
||||
cascadeTracker.setRollback(false);
|
||||
cascadeTracker.invalidateState();
|
||||
cascadeTracker.ifPresent(c -> c.setRollback(false));
|
||||
cascadeTracker.ifPresent(c -> invalidateState());
|
||||
|
||||
//Release the locks
|
||||
contentTracker.getWriteLock().release();
|
||||
cascadeTracker.getWriteLock().release();
|
||||
cascadeTracker.ifPresent(c -> c.getWriteLock().release());
|
||||
|
||||
rollbackCount.incrementAndGet();
|
||||
}
|
||||
|
@@ -32,6 +32,7 @@ import org.alfresco.repo.index.shard.ShardState;
|
||||
import org.alfresco.solr.BoundedDeque;
|
||||
import org.alfresco.solr.InformationServer;
|
||||
import org.alfresco.solr.NodeReport;
|
||||
import org.alfresco.solr.SolrInformationServer;
|
||||
import org.alfresco.solr.TrackerState;
|
||||
import org.alfresco.solr.adapters.IOpenBitSet;
|
||||
import org.alfresco.solr.client.GetNodesParameters;
|
||||
@@ -83,6 +84,8 @@ public class MetadataTracker extends CoreStatePublisher implements Tracker
|
||||
* {@link org.alfresco.solr.client.SOLRAPIClient#GET_TX_INTERVAL_COMMIT_TIME}
|
||||
*/
|
||||
private boolean txIntervalCommitTimeServiceAvailable = false;
|
||||
/** Whether the cascade tracking is enabled. */
|
||||
private boolean cascadeTrackerEnabled = true;
|
||||
|
||||
public MetadataTracker(final boolean isMaster, Properties p, SOLRAPIClient client, String coreName,
|
||||
InformationServer informationServer)
|
||||
@@ -107,6 +110,7 @@ public class MetadataTracker extends CoreStatePublisher implements Tracker
|
||||
transactionDocsBatchSize = Integer.parseInt(p.getProperty("alfresco.transactionDocsBatchSize", "100"));
|
||||
nodeBatchSize = Integer.parseInt(p.getProperty("alfresco.nodeBatchSize", "10"));
|
||||
threadHandler = new ThreadHandler(p, coreName, "MetadataTracker");
|
||||
cascadeTrackerEnabled = informationServer.cascadeTrackingEnabled();
|
||||
|
||||
// In order to apply performance optimizations, checking the availability of Repo Web Scripts is required.
|
||||
// As these services are available from ACS 6.2
|
||||
@@ -957,7 +961,7 @@ public class MetadataTracker extends CoreStatePublisher implements Tracker
|
||||
List<Node> filteredNodes = filterNodes(nodes);
|
||||
if(filteredNodes.size() > 0)
|
||||
{
|
||||
this.infoServer.indexNodes(filteredNodes, true, false);
|
||||
this.infoServer.indexNodes(filteredNodes, true);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -977,9 +981,8 @@ public class MetadataTracker extends CoreStatePublisher implements Tracker
|
||||
{
|
||||
filteredList.add(node);
|
||||
}
|
||||
else
|
||||
else if (cascadeTrackerEnabled)
|
||||
{
|
||||
|
||||
if(node.getStatus() == SolrApiNodeStatus.UPDATED)
|
||||
{
|
||||
Node doCascade = new Node();
|
||||
|
@@ -30,4 +30,7 @@ alfresco.cross.locale.property.1={http://www.alfresco.org/model/content/1.0}lock
|
||||
# alfresco.cross.locale.datatype.1={http://www.alfresco.org/model/dictionary/1.0}content
|
||||
# alfresco.cross.locale.datatype.2={http://www.alfresco.org/model/dictionary/1.0}mltext
|
||||
|
||||
alfresco.model.tracker.cron=0/10 * * * * ? *
|
||||
alfresco.model.tracker.cron=0/10 * * * * ? *
|
||||
|
||||
# Whether path queries are enabled.
|
||||
alfresco.cascade.tracker.enabled=true
|
||||
|
@@ -18,6 +18,30 @@
|
||||
*/
|
||||
package org.alfresco.solr.lifecycle;
|
||||
|
||||
import static java.util.Arrays.asList;
|
||||
|
||||
import static org.alfresco.solr.SolrInformationServer.CASCADE_TRACKER_ENABLED;
|
||||
import static org.alfresco.solr.tracker.Tracker.Type.ACL;
|
||||
import static org.alfresco.solr.tracker.Tracker.Type.CASCADE;
|
||||
import static org.alfresco.solr.tracker.Tracker.Type.CONTENT;
|
||||
import static org.alfresco.solr.tracker.Tracker.Type.METADATA;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.mockito.ArgumentMatchers.any;
|
||||
import static org.mockito.ArgumentMatchers.eq;
|
||||
import static org.mockito.ArgumentMatchers.same;
|
||||
import static org.mockito.Mockito.mock;
|
||||
import static org.mockito.Mockito.never;
|
||||
import static org.mockito.Mockito.verify;
|
||||
import static org.mockito.Mockito.when;
|
||||
import static org.mockito.MockitoAnnotations.initMocks;
|
||||
|
||||
import java.util.List;
|
||||
import java.util.Properties;
|
||||
import java.util.Set;
|
||||
import java.util.stream.Collectors;
|
||||
|
||||
import org.alfresco.solr.SolrInformationServer;
|
||||
import org.alfresco.solr.client.SOLRAPIClient;
|
||||
import org.alfresco.solr.tracker.AclTracker;
|
||||
@@ -26,6 +50,7 @@ import org.alfresco.solr.tracker.ContentTracker;
|
||||
import org.alfresco.solr.tracker.MetadataTracker;
|
||||
import org.alfresco.solr.tracker.SolrTrackerScheduler;
|
||||
import org.alfresco.solr.tracker.Tracker;
|
||||
import org.alfresco.solr.tracker.Tracker.Type;
|
||||
import org.alfresco.solr.tracker.TrackerRegistry;
|
||||
import org.apache.solr.core.SolrConfig;
|
||||
import org.apache.solr.core.SolrCore;
|
||||
@@ -36,20 +61,6 @@ import org.mockito.Mock;
|
||||
import org.mockito.junit.MockitoJUnitRunner;
|
||||
import org.xml.sax.InputSource;
|
||||
|
||||
import java.util.List;
|
||||
import java.util.Properties;
|
||||
|
||||
import static java.util.Arrays.asList;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.mockito.ArgumentMatchers.any;
|
||||
import static org.mockito.ArgumentMatchers.eq;
|
||||
import static org.mockito.ArgumentMatchers.same;
|
||||
import static org.mockito.Mockito.mock;
|
||||
import static org.mockito.Mockito.verify;
|
||||
import static org.mockito.Mockito.when;
|
||||
|
||||
/**
|
||||
* Unit tests for the {@link SolrCoreLoadListener}.
|
||||
*
|
||||
@@ -83,6 +94,8 @@ public class SolrCoreLoadListenerTest
|
||||
@Before
|
||||
public void setUp()
|
||||
{
|
||||
initMocks(this);
|
||||
|
||||
listener = new SolrCoreLoadListener(core);
|
||||
when(core.getName()).thenReturn(coreName);
|
||||
|
||||
@@ -104,7 +117,29 @@ public class SolrCoreLoadListenerTest
|
||||
verify(scheduler).schedule(any(MetadataTracker.class), eq(coreName), same(coreProperties));
|
||||
verify(scheduler).schedule(any(CascadeTracker.class), eq(coreName), same(coreProperties));
|
||||
|
||||
assertEquals(4, coreTrackers.size());
|
||||
Set<Type> trackerTypes = coreTrackers.stream().map(Tracker::getType).collect(Collectors.toSet());
|
||||
assertEquals("Unexpected trackers found.", Set.of(ACL, CONTENT, METADATA, CASCADE), trackerTypes);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testDisabledCascadeTracking()
|
||||
{
|
||||
coreProperties.put(CASCADE_TRACKER_ENABLED, "false");
|
||||
|
||||
List<Tracker> coreTrackers = listener.createAndScheduleCoreTrackers(core, registry, coreProperties, scheduler, api, informationServer);
|
||||
|
||||
verify(registry).register(eq(coreName), any(AclTracker.class));
|
||||
verify(registry).register(eq(coreName), any(ContentTracker.class));
|
||||
verify(registry).register(eq(coreName), any(MetadataTracker.class));
|
||||
verify(registry, never()).register(eq(coreName), any(CascadeTracker.class));
|
||||
|
||||
verify(scheduler).schedule(any(AclTracker.class), eq(coreName), same(coreProperties));
|
||||
verify(scheduler).schedule(any(ContentTracker.class), eq(coreName), same(coreProperties));
|
||||
verify(scheduler).schedule(any(MetadataTracker.class), eq(coreName), same(coreProperties));
|
||||
verify(scheduler, never()).schedule(any(CascadeTracker.class), eq(coreName), same(coreProperties));
|
||||
|
||||
Set<Type> trackerTypes = coreTrackers.stream().map(Tracker::getType).collect(Collectors.toSet());
|
||||
assertEquals("Unexpected trackers found.", Set.of(ACL, CONTENT, METADATA), trackerTypes);
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@@ -52,7 +52,7 @@ import java.util.List;
|
||||
|
||||
@LuceneTestCase.SuppressCodecs({"Appending","Lucene3x","Lucene40","Lucene41","Lucene42","Lucene43", "Lucene44", "Lucene45","Lucene46","Lucene47","Lucene48","Lucene49"})
|
||||
@SolrTestCaseJ4.SuppressSSL
|
||||
public class CascadeTrackerIT extends AbstractAlfrescoSolrIT
|
||||
public class CascadingIT extends AbstractAlfrescoSolrIT
|
||||
{
|
||||
private static long MAX_WAIT_TIME = 80000;
|
||||
|
@@ -55,7 +55,7 @@ import java.util.Properties;
|
||||
*/
|
||||
@SolrTestCaseJ4.SuppressSSL
|
||||
@LuceneTestCase.SuppressCodecs({"Appending","Lucene3x","Lucene40","Lucene41","Lucene42","Lucene43", "Lucene44", "Lucene45","Lucene46","Lucene47","Lucene48","Lucene49"})
|
||||
public class DistributedCascadeTrackerIT extends AbstractAlfrescoDistributedIT
|
||||
public class DistributedCascadeIT extends AbstractAlfrescoDistributedIT
|
||||
{
|
||||
private Node parentFolder;
|
||||
private NodeMetaData parentFolderMetadata;
|
@@ -120,7 +120,7 @@ public class MetadataTrackerTest
|
||||
this.metadataTracker.doTrack();
|
||||
|
||||
InOrder inOrder = inOrder(srv);
|
||||
inOrder.verify(srv).indexNodes(nodes, true, false);
|
||||
inOrder.verify(srv).indexNodes(nodes, true);
|
||||
inOrder.verify(srv).indexTransaction(tx, true);
|
||||
inOrder.verify(srv).commit();
|
||||
}
|
||||
|
Reference in New Issue
Block a user