mirror of
https://github.com/Alfresco/alfresco-community-repo.git
synced 2025-07-24 17:32:48 +00:00
Merge from HEAD to WCM-DEV2.
git-svn-id: https://svn.alfresco.com/repos/alfresco-enterprise/alfresco/BRANCHES/WCM-DEV2/root@3659 c4b6b30b-aa2e-2d43-bbcb-ca4b014f7261
This commit is contained in:
@@ -1630,10 +1630,7 @@ public abstract class BaseNodeServiceTest extends BaseSpringTest
|
||||
NodeRef defRef = pathDefRef.getChildRef();
|
||||
|
||||
// now browse down using the node service
|
||||
NodeRef checkParentRef = nodeService.getChildByName(rootNodeRef, ASSOC_TYPE_QNAME_TEST_CHILDREN, parentRef.getId());
|
||||
assertNotNull("First level, non-named node not found", checkParentRef);
|
||||
assertEquals(parentRef, checkParentRef);
|
||||
NodeRef checkAbcRef = nodeService.getChildByName(checkParentRef, ASSOC_TYPE_QNAME_TEST_CONTAINS, "abc");
|
||||
NodeRef checkAbcRef = nodeService.getChildByName(parentRef, ASSOC_TYPE_QNAME_TEST_CONTAINS, "abc");
|
||||
assertNotNull("Second level, named node 'ABC' not found", checkAbcRef);
|
||||
assertEquals(abcRef, checkAbcRef);
|
||||
NodeRef checkDefRef = nodeService.getChildByName(checkAbcRef, ASSOC_TYPE_QNAME_TEST_CONTAINS, "def");
|
||||
|
@@ -149,7 +149,7 @@ public class DbNodeServiceImpl extends AbstractNodeServiceImpl
|
||||
else
|
||||
{
|
||||
return new NodeRef.Status(
|
||||
nodeStatus.getChangeTxnId(),
|
||||
nodeStatus.getTransaction().getChangeTxnId(),
|
||||
nodeStatus.isDeleted());
|
||||
}
|
||||
}
|
||||
@@ -1440,11 +1440,11 @@ public class DbNodeServiceImpl extends AbstractNodeServiceImpl
|
||||
// update old status
|
||||
NodeStatus oldNodeStatus = nodeDaoService.getNodeStatus(oldNodeRef, true);
|
||||
oldNodeStatus.setNode(null);
|
||||
oldNodeStatus.setChangeTxnId(txnId);
|
||||
oldNodeStatus.getTransaction().setChangeTxnId(txnId);
|
||||
// create the new status
|
||||
NodeStatus newNodeStatus = nodeDaoService.getNodeStatus(newNodeRef, true);
|
||||
newNodeStatus.setNode(nodeToMove);
|
||||
newNodeStatus.setChangeTxnId(txnId);
|
||||
newNodeStatus.getTransaction().setChangeTxnId(txnId);
|
||||
}
|
||||
}
|
||||
|
||||
|
@@ -16,9 +16,14 @@
|
||||
*/
|
||||
package org.alfresco.repo.node.db.hibernate;
|
||||
|
||||
import java.io.Serializable;
|
||||
import java.net.InetAddress;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collection;
|
||||
import java.util.List;
|
||||
import java.util.concurrent.locks.ReentrantReadWriteLock;
|
||||
import java.util.concurrent.locks.ReentrantReadWriteLock.ReadLock;
|
||||
import java.util.concurrent.locks.ReentrantReadWriteLock.WriteLock;
|
||||
import java.util.zip.CRC32;
|
||||
|
||||
import org.alfresco.error.AlfrescoRuntimeException;
|
||||
@@ -28,13 +33,17 @@ import org.alfresco.repo.domain.Node;
|
||||
import org.alfresco.repo.domain.NodeAssoc;
|
||||
import org.alfresco.repo.domain.NodeKey;
|
||||
import org.alfresco.repo.domain.NodeStatus;
|
||||
import org.alfresco.repo.domain.Server;
|
||||
import org.alfresco.repo.domain.Store;
|
||||
import org.alfresco.repo.domain.StoreKey;
|
||||
import org.alfresco.repo.domain.Transaction;
|
||||
import org.alfresco.repo.domain.hibernate.ChildAssocImpl;
|
||||
import org.alfresco.repo.domain.hibernate.NodeAssocImpl;
|
||||
import org.alfresco.repo.domain.hibernate.NodeImpl;
|
||||
import org.alfresco.repo.domain.hibernate.NodeStatusImpl;
|
||||
import org.alfresco.repo.domain.hibernate.ServerImpl;
|
||||
import org.alfresco.repo.domain.hibernate.StoreImpl;
|
||||
import org.alfresco.repo.domain.hibernate.TransactionImpl;
|
||||
import org.alfresco.repo.node.db.NodeDaoService;
|
||||
import org.alfresco.repo.transaction.AlfrescoTransactionSupport;
|
||||
import org.alfresco.repo.transaction.TransactionalDao;
|
||||
@@ -71,9 +80,14 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
private static final String QUERY_GET_TARGET_ASSOCS = "node.GetTargetAssocs";
|
||||
private static final String QUERY_GET_SOURCE_ASSOCS = "node.GetSourceAssocs";
|
||||
private static final String QUERY_GET_CONTENT_DATA_STRINGS = "node.GetContentDataStrings";
|
||||
private static final String QUERY_GET_SERVER_BY_IPADDRESS = "server.getServerByIpAddress";
|
||||
|
||||
/** a uuid identifying this unique instance */
|
||||
private String uuid;
|
||||
private final String uuid;
|
||||
|
||||
private final ReadLock serverReadLock;
|
||||
private final WriteLock serverWriteLock;
|
||||
private Server server;
|
||||
|
||||
/**
|
||||
*
|
||||
@@ -81,6 +95,10 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
public HibernateNodeDaoServiceImpl()
|
||||
{
|
||||
this.uuid = GUID.generate();
|
||||
|
||||
ReentrantReadWriteLock serverReadWriteLock = new ReentrantReadWriteLock();
|
||||
serverReadLock = serverReadWriteLock.readLock();
|
||||
serverWriteLock = serverReadWriteLock.writeLock();
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -108,6 +126,93 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
return uuid.hashCode();
|
||||
}
|
||||
|
||||
/**
|
||||
* Gets/creates the <b>server</b> instance to use for the life of this instance
|
||||
*/
|
||||
private Server getServer()
|
||||
{
|
||||
// get readlock
|
||||
serverReadLock.lock();
|
||||
try
|
||||
{
|
||||
if (server != null)
|
||||
{
|
||||
return server;
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
serverReadLock.unlock();
|
||||
}
|
||||
// get the write lock
|
||||
serverWriteLock.lock();
|
||||
try
|
||||
{
|
||||
final String ipAddress = InetAddress.getLocalHost().getHostAddress();
|
||||
HibernateCallback callback = new HibernateCallback()
|
||||
{
|
||||
public Object doInHibernate(Session session)
|
||||
{
|
||||
Query query = session
|
||||
.getNamedQuery(HibernateNodeDaoServiceImpl.QUERY_GET_SERVER_BY_IPADDRESS)
|
||||
.setString("ipAddress", ipAddress);
|
||||
return query.uniqueResult();
|
||||
}
|
||||
};
|
||||
server = (Server) getHibernateTemplate().execute(callback);
|
||||
// create it if it doesn't exist
|
||||
if (server == null)
|
||||
{
|
||||
server = new ServerImpl();
|
||||
server.setIpAddress(ipAddress);
|
||||
try
|
||||
{
|
||||
getSession().save(server);
|
||||
}
|
||||
catch (DataIntegrityViolationException e)
|
||||
{
|
||||
// get it again
|
||||
server = (Server) getHibernateTemplate().execute(callback);
|
||||
if (server == null)
|
||||
{
|
||||
throw new AlfrescoRuntimeException("Unable to create server instance: " + ipAddress);
|
||||
}
|
||||
}
|
||||
}
|
||||
return server;
|
||||
}
|
||||
catch (Exception e)
|
||||
{
|
||||
throw new AlfrescoRuntimeException("Failed to create server instance", e);
|
||||
}
|
||||
finally
|
||||
{
|
||||
serverWriteLock.unlock();
|
||||
}
|
||||
}
|
||||
|
||||
private static final String RESOURCE_KEY_TRANSACTION_ID = "hibernate.transaction.id";
|
||||
private Transaction getCurrentTransaction()
|
||||
{
|
||||
Transaction transaction = null;
|
||||
Serializable txnId = (Serializable) AlfrescoTransactionSupport.getResource(RESOURCE_KEY_TRANSACTION_ID);
|
||||
if (txnId == null)
|
||||
{
|
||||
// no transaction instance has been bound to the transaction
|
||||
transaction = new TransactionImpl();
|
||||
transaction.setChangeTxnId(AlfrescoTransactionSupport.getTransactionId());
|
||||
transaction.setServer(getServer());
|
||||
txnId = getHibernateTemplate().save(transaction);
|
||||
// bind the id
|
||||
AlfrescoTransactionSupport.bindResource(RESOURCE_KEY_TRANSACTION_ID, txnId);
|
||||
}
|
||||
else
|
||||
{
|
||||
transaction = (Transaction) getHibernateTemplate().get(TransactionImpl.class, txnId);
|
||||
}
|
||||
return transaction;
|
||||
}
|
||||
|
||||
/**
|
||||
* Does this <tt>Session</tt> contain any changes which must be
|
||||
* synchronized with the store?
|
||||
@@ -218,7 +323,7 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
{
|
||||
status = new NodeStatusImpl();
|
||||
status.setKey(nodeKey);
|
||||
status.setChangeTxnId(AlfrescoTransactionSupport.getTransactionId());
|
||||
status.setTransaction(getCurrentTransaction());
|
||||
getHibernateTemplate().save(status);
|
||||
}
|
||||
// done
|
||||
@@ -237,7 +342,7 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
}
|
||||
else
|
||||
{
|
||||
status.setChangeTxnId(AlfrescoTransactionSupport.getTransactionId());
|
||||
status.getTransaction().setChangeTxnId(AlfrescoTransactionSupport.getTransactionId());
|
||||
}
|
||||
}
|
||||
|
||||
@@ -259,13 +364,13 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
// If that is the case, then the session has to be flushed so that the database
|
||||
// constraints aren't violated as the node creation will write to the database to
|
||||
// get an ID
|
||||
if (status.getChangeTxnId().equals(AlfrescoTransactionSupport.getTransactionId()))
|
||||
if (status.getTransaction().getChangeTxnId().equals(AlfrescoTransactionSupport.getTransactionId()))
|
||||
{
|
||||
// flush
|
||||
getHibernateTemplate().flush();
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// build a concrete node based on a bootstrap type
|
||||
Node node = new NodeImpl();
|
||||
// set other required properties
|
||||
@@ -277,7 +382,11 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
|
||||
// set required status properties
|
||||
status.setNode(node);
|
||||
status.setChangeTxnId(AlfrescoTransactionSupport.getTransactionId());
|
||||
// assign a transaction
|
||||
if (status.getTransaction() == null)
|
||||
{
|
||||
status.setTransaction(getCurrentTransaction());
|
||||
}
|
||||
// persist the nodestatus
|
||||
getHibernateTemplate().save(status);
|
||||
|
||||
@@ -331,7 +440,7 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
NodeRef nodeRef = node.getNodeRef();
|
||||
NodeStatus nodeStatus = getNodeStatus(nodeRef, true);
|
||||
nodeStatus.setNode(null);
|
||||
nodeStatus.setChangeTxnId(AlfrescoTransactionSupport.getTransactionId());
|
||||
nodeStatus.getTransaction().setChangeTxnId(AlfrescoTransactionSupport.getTransactionId());
|
||||
// finally delete the node
|
||||
getHibernateTemplate().delete(node);
|
||||
// flush to ensure constraints can't be violated
|
||||
@@ -371,7 +480,7 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
{
|
||||
/*
|
||||
* This initial child association creation will fail IFF there is already
|
||||
* an association of the given type between the two nodes. For new association
|
||||
* an association of the given type and name between the two nodes. For new association
|
||||
* creation, this can only occur if two transactions attempt to create a secondary
|
||||
* child association between the same two nodes. As this is unlikely, it is
|
||||
* appropriate to just throw a runtime exception and let the second transaction
|
||||
@@ -383,28 +492,18 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
* if the association is recreated subsequently.
|
||||
*/
|
||||
|
||||
String uuid = childNode.getUuid();
|
||||
// assign a random name to the node
|
||||
String randomName = GUID.generate();
|
||||
|
||||
ChildAssoc assoc = new ChildAssocImpl();
|
||||
assoc.setTypeQName(assocTypeQName);
|
||||
assoc.setChildNodeName(getShortName(uuid));
|
||||
assoc.setChildNodeNameCrc(getCrc(uuid));
|
||||
assoc.setChildNodeName(randomName);
|
||||
assoc.setChildNodeNameCrc(-1L); // random names compete only with each other
|
||||
assoc.setQname(qname);
|
||||
assoc.setIsPrimary(isPrimary);
|
||||
assoc.buildAssociation(parentNode, childNode);
|
||||
// persist it, catching the duplicate child name
|
||||
try
|
||||
{
|
||||
getHibernateTemplate().save(assoc);
|
||||
}
|
||||
catch (DataIntegrityViolationException e)
|
||||
{
|
||||
throw new AlfrescoRuntimeException("An association already exists between the two nodes: \n" +
|
||||
" parent: " + parentNode.getId() + "\n" +
|
||||
" child: " + childNode.getId() + "\n" +
|
||||
" assoc: " + assocTypeQName,
|
||||
e);
|
||||
}
|
||||
getHibernateTemplate().save(assoc);
|
||||
// done
|
||||
return assoc;
|
||||
}
|
||||
@@ -422,17 +521,22 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
|
||||
*/
|
||||
|
||||
String childNameNew = null;
|
||||
long crc = -1;
|
||||
if (childName == null)
|
||||
{
|
||||
childNameNew = childAssoc.getChild().getUuid();
|
||||
// random names compete only with each other, i.e. not at all
|
||||
childNameNew = GUID.generate();
|
||||
crc = -1;
|
||||
}
|
||||
else
|
||||
{
|
||||
// assigned names compete exactly
|
||||
childNameNew = childName.toLowerCase();
|
||||
crc = getCrc(childNameNew);
|
||||
}
|
||||
|
||||
final String childNameNewShort = getShortName(childNameNew);
|
||||
final long childNameNewCrc = getCrc(childNameNew);
|
||||
final long childNameNewCrc = crc;
|
||||
|
||||
// check if the name has changed
|
||||
if (childAssoc.getChildNodeNameCrc() == childNameNewCrc)
|
||||
|
@@ -62,13 +62,13 @@ import org.springframework.orm.hibernate3.support.HibernateDaoSupport;
|
||||
* database is static then the L2 cache usage can be set to use
|
||||
* the <code>NORMAL</code> mode. <code>REFRESH</code> should be
|
||||
* used where the server will still be accessed from some clients
|
||||
* despite the database changing.
|
||||
* despite the database changing. <code>NORMAL</code> can be used
|
||||
* in the case of the caches being clustered, i.e. the caches will
|
||||
* not be out of date w.r.t. the database.
|
||||
* </li>
|
||||
* <li>
|
||||
* This process should not run continuously on a live
|
||||
* server as it would be performing unecessary work.
|
||||
* If it was left running, however, it would not
|
||||
* lead to data corruption or such-like. Use the
|
||||
* This process should only be used continuously where the index
|
||||
* transactions are following the database transactions. Use the
|
||||
* {@link #setRunContinuously(boolean) runContinuously} property
|
||||
* to change this behaviour.
|
||||
* </li>
|
||||
@@ -91,7 +91,7 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
private static boolean started = false;
|
||||
/** The current transaction ID being processed */
|
||||
private static String currentTxnId = START_TXN_ID;
|
||||
/** kept to notify the thread that it should quite */
|
||||
/** kept to notify the thread that it should quit */
|
||||
private boolean killThread = false;
|
||||
|
||||
/** provides transactions to atomically index each missed transaction */
|
||||
@@ -104,8 +104,6 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
private SearchService searcher;
|
||||
/** the component giving direct access to <b>node</b> instances */
|
||||
private NodeService nodeService;
|
||||
/** the stores to reindex */
|
||||
private List<StoreRef> storeRefs;
|
||||
/** set this to run the index recovery component */
|
||||
private boolean executeFullRecovery;
|
||||
/** set this on to keep checking for new transactions and never stop */
|
||||
@@ -125,8 +123,6 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
|
||||
public FullIndexRecoveryComponent()
|
||||
{
|
||||
this.storeRefs = new ArrayList<StoreRef>(2);
|
||||
|
||||
this.killThread = false;
|
||||
this.executeFullRecovery = false;
|
||||
this.runContinuously = false;
|
||||
@@ -193,21 +189,6 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
this.nodeService = nodeService;
|
||||
}
|
||||
|
||||
/**
|
||||
* Set the stores that need reindexing
|
||||
*
|
||||
* @param storeRefStrings a list of strings representing store references
|
||||
*/
|
||||
public void setStores(List<String> storeRefStrings)
|
||||
{
|
||||
storeRefs.clear();
|
||||
for (String storeRefStr : storeRefStrings)
|
||||
{
|
||||
StoreRef storeRef = new StoreRef(storeRefStr);
|
||||
storeRefs.add(storeRef);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Set this to <code>true</code> to initiate the full index recovery.
|
||||
* <p>
|
||||
@@ -299,6 +280,7 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
{
|
||||
public Object doWork()
|
||||
{
|
||||
List<StoreRef> storeRefs = nodeService.getStores();
|
||||
// reindex each store
|
||||
for (StoreRef storeRef : storeRefs)
|
||||
{
|
||||
@@ -352,8 +334,7 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
if (logger.isDebugEnabled())
|
||||
{
|
||||
logger.debug("Full index recovery thread started: \n" +
|
||||
" continuous: " + runContinuously + "\n" +
|
||||
" stores: " + storeRefs);
|
||||
" continuous: " + runContinuously);
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -377,8 +358,8 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
// reindex nodes
|
||||
List<String> txnsIndexed = FullIndexRecoveryComponent.this.reindexNodes();
|
||||
// reindex missing content
|
||||
@SuppressWarnings("unused")
|
||||
int missingContentCount = FullIndexRecoveryComponent.this.reindexMissingContent();
|
||||
// @SuppressWarnings("unused")
|
||||
// int missingContentCount = FullIndexRecoveryComponent.this.reindexMissingContent();
|
||||
// check if the process should terminate
|
||||
if (txnsIndexed.size() == 0 && !runContinuously)
|
||||
{
|
||||
@@ -417,73 +398,6 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* @return Returns the number of documents reindexed
|
||||
*/
|
||||
private int reindexMissingContent()
|
||||
{
|
||||
int count = 0;
|
||||
for (StoreRef storeRef : storeRefs)
|
||||
{
|
||||
count += reindexMissingContent(storeRef);
|
||||
}
|
||||
return count;
|
||||
}
|
||||
|
||||
/**
|
||||
* @param storeRef the store to check for missing content
|
||||
* @return Returns the number of documents reindexed
|
||||
*/
|
||||
private int reindexMissingContent(StoreRef storeRef)
|
||||
{
|
||||
SearchParameters sp = new SearchParameters();
|
||||
sp.addStore(storeRef);
|
||||
|
||||
// search for it in the index
|
||||
String query = "TEXT:" + LuceneIndexerImpl.NOT_INDEXED_CONTENT_MISSING;
|
||||
sp.setLanguage(SearchService.LANGUAGE_LUCENE);
|
||||
sp.setQuery(query);
|
||||
ResultSet results = null;
|
||||
try
|
||||
{
|
||||
results = searcher.query(sp);
|
||||
|
||||
int count = 0;
|
||||
// loop over the results and get the details of the nodes that have missing content
|
||||
List<ChildAssociationRef> assocRefs = results.getChildAssocRefs();
|
||||
for (ChildAssociationRef assocRef : assocRefs)
|
||||
{
|
||||
final NodeRef childNodeRef = assocRef.getChildRef();
|
||||
// prompt for a reindex - it might fail again, but we just keep plugging away
|
||||
TransactionWork<Object> reindexWork = new TransactionWork<Object>()
|
||||
{
|
||||
public Object doWork()
|
||||
{
|
||||
indexer.updateNode(childNodeRef);
|
||||
return null;
|
||||
}
|
||||
};
|
||||
TransactionUtil.executeInNonPropagatingUserTransaction(transactionService, reindexWork);
|
||||
count++;
|
||||
}
|
||||
// done
|
||||
if (logger.isDebugEnabled())
|
||||
{
|
||||
logger.debug("Reindexed missing content: \n" +
|
||||
" store: " + storeRef + "\n" +
|
||||
" node count: " + count);
|
||||
}
|
||||
return count;
|
||||
}
|
||||
finally
|
||||
{
|
||||
if (results != null)
|
||||
{
|
||||
results.close();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* @return Returns the transaction ID just reindexed, i.e. where some work was performed
|
||||
*/
|
||||
@@ -572,16 +486,16 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
getSession().setCacheMode(l2CacheMode);
|
||||
|
||||
// reindex each store
|
||||
for (StoreRef storeRef : storeRefs)
|
||||
{
|
||||
if (!nodeService.exists(storeRef))
|
||||
{
|
||||
// the store is not present
|
||||
continue;
|
||||
}
|
||||
// reindex for store
|
||||
reindexNodes(storeRef, changeTxnId);
|
||||
}
|
||||
// for (StoreRef storeRef : storeRefs)
|
||||
// {
|
||||
// if (!nodeService.exists(storeRef))
|
||||
// {
|
||||
// // the store is not present
|
||||
// continue;
|
||||
// }
|
||||
// // reindex for store
|
||||
// reindexNodes(storeRef, changeTxnId);
|
||||
// }
|
||||
// done
|
||||
return null;
|
||||
}
|
||||
@@ -675,10 +589,10 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
};
|
||||
|
||||
/**
|
||||
* Retrieve all transaction IDs that are greater than the given transaction ID.
|
||||
* Retrieve next 50 transaction IDs that are greater than the given transaction ID.
|
||||
*
|
||||
* @param currentTxnId the transaction ID that must be less than all returned results
|
||||
* @return Returns an ordered list of transaction IDs
|
||||
* @return Returns an ordered list of the next 50 transaction IDs
|
||||
*/
|
||||
@SuppressWarnings("unchecked")
|
||||
public List<String> getNextChangeTxnIds(final String currentTxnId)
|
||||
@@ -689,6 +603,7 @@ public class FullIndexRecoveryComponent extends HibernateDaoSupport implements I
|
||||
{
|
||||
Query query = session.getNamedQuery(QUERY_GET_NEXT_CHANGE_TXN_IDS);
|
||||
query.setString("currentTxnId", currentTxnId)
|
||||
.setMaxResults(50)
|
||||
.setReadOnly(true);
|
||||
return query.list();
|
||||
}
|
||||
|
@@ -123,7 +123,7 @@ public class FullIndexRecoveryComponentTest extends TestCase
|
||||
String txnId = TransactionUtil.executeInNonPropagatingUserTransaction(txnService, dropNodeIndexWork);
|
||||
|
||||
indexRecoverer.setExecuteFullRecovery(true);
|
||||
indexRecoverer.setStores(storeRefStrings);
|
||||
// indexRecoverer.setStores(storeRefStrings);
|
||||
// reindex
|
||||
indexRecoverer.reindex();
|
||||
|
||||
|
@@ -0,0 +1,741 @@
|
||||
///*
|
||||
// * Copyright (C) 2005-2006 Alfresco, Inc.
|
||||
// *
|
||||
// * Licensed under the Mozilla Public License version 1.1
|
||||
// * with a permitted attribution clause. You may obtain a
|
||||
// * copy of the License at
|
||||
// *
|
||||
// * http://www.alfresco.org/legal/license.txt
|
||||
// *
|
||||
// * Unless required by applicable law or agreed to in writing,
|
||||
// * software distributed under the License is distributed on an
|
||||
// * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
|
||||
// * either express or implied. See the License for the specific
|
||||
// * language governing permissions and limitations under the
|
||||
// * License.
|
||||
// */
|
||||
//package org.alfresco.repo.node.index;
|
||||
//
|
||||
//import java.util.ArrayList;
|
||||
//import java.util.List;
|
||||
//
|
||||
//import org.alfresco.error.AlfrescoRuntimeException;
|
||||
//import org.alfresco.model.ContentModel;
|
||||
//import org.alfresco.repo.domain.NodeStatus;
|
||||
//import org.alfresco.repo.search.Indexer;
|
||||
//import org.alfresco.repo.search.impl.lucene.LuceneIndexerImpl;
|
||||
//import org.alfresco.repo.search.impl.lucene.fts.FullTextSearchIndexer;
|
||||
//import org.alfresco.repo.transaction.TransactionUtil;
|
||||
//import org.alfresco.repo.transaction.TransactionUtil.TransactionWork;
|
||||
//import org.alfresco.service.cmr.repository.ChildAssociationRef;
|
||||
//import org.alfresco.service.cmr.repository.NodeRef;
|
||||
//import org.alfresco.service.cmr.repository.NodeService;
|
||||
//import org.alfresco.service.cmr.repository.StoreRef;
|
||||
//import org.alfresco.service.cmr.search.ResultSet;
|
||||
//import org.alfresco.service.cmr.search.SearchParameters;
|
||||
//import org.alfresco.service.cmr.search.SearchService;
|
||||
//import org.alfresco.service.transaction.TransactionService;
|
||||
//import org.apache.commons.logging.Log;
|
||||
//import org.apache.commons.logging.LogFactory;
|
||||
//import org.hibernate.CacheMode;
|
||||
//import org.hibernate.Query;
|
||||
//import org.hibernate.Session;
|
||||
//import org.springframework.orm.hibernate3.HibernateCallback;
|
||||
//import org.springframework.orm.hibernate3.support.HibernateDaoSupport;
|
||||
//
|
||||
///**
|
||||
// * Ensures that the FTS indexing picks up on any outstanding documents that
|
||||
// * require indexing.
|
||||
// * <p>
|
||||
// * This component must be used as a singleton (one per VM) and may only be
|
||||
// * called to reindex once. It will start a thread that processes all available
|
||||
// * transactions and keeps checking to ensure that the index is up to date with
|
||||
// * the latest database changes.
|
||||
// * <p>
|
||||
// * <b>The following points are important:</b>
|
||||
// * <ul>
|
||||
// * <li>
|
||||
// * By default, the Hibernate L2 cache is used during processing.
|
||||
// * This can be disabled by either disabling the L2 cache globally
|
||||
// * for the server (not recommended) or by setting the
|
||||
// * {@link #setL2CacheMode(String) l2CacheMode} property. If the
|
||||
// * database is static then the L2 cache usage can be set to use
|
||||
// * the <code>NORMAL</code> mode. <code>REFRESH</code> should be
|
||||
// * used where the server will still be accessed from some clients
|
||||
// * despite the database changing. <code>NORMAL</code> can be used
|
||||
// * in the case of the caches being clustered, i.e. the caches will
|
||||
// * not be out of date w.r.t. the database.
|
||||
// * </li>
|
||||
// * <li>
|
||||
// * This process should only be used continuously where the index
|
||||
// * transactions are following the database transactions. Use the
|
||||
// * {@link #setRunContinuously(boolean) runContinuously} property
|
||||
// * to change this behaviour.
|
||||
// * </li>
|
||||
// * </ul>
|
||||
// *
|
||||
// * @author Derek Hulley
|
||||
// */
|
||||
//public class MissingContentReindexComponent extends HibernateDaoSupport implements IndexRecovery
|
||||
//{
|
||||
// public static final String QUERY_GET_NEXT_CHANGE_TXN_IDS = "node.GetNextChangeTxnIds";
|
||||
// public static final String QUERY_GET_CHANGED_NODE_STATUSES = "node.GetChangedNodeStatuses";
|
||||
// public static final String QUERY_GET_DELETED_NODE_STATUSES = "node.GetDeletedNodeStatuses";
|
||||
// public static final String QUERY_GET_CHANGED_NODE_STATUSES_COUNT = "node.GetChangedNodeStatusesCount";
|
||||
//
|
||||
// private static final String START_TXN_ID = "000";
|
||||
//
|
||||
// private static Log logger = LogFactory.getLog(FullIndexRecoveryComponent.class);
|
||||
//
|
||||
// /** ensures that this process is kicked off once per VM */
|
||||
// private static boolean started = false;
|
||||
// /** The current transaction ID being processed */
|
||||
// private static String currentTxnId = START_TXN_ID;
|
||||
// /** kept to notify the thread that it should quite */
|
||||
// private boolean killThread = false;
|
||||
//
|
||||
// /** provides transactions to atomically index each missed transaction */
|
||||
// private TransactionService transactionService;
|
||||
// /** the component to index the node hierarchy */
|
||||
// private Indexer indexer;
|
||||
// /** the FTS indexer that we will prompt to pick up on any un-indexed text */
|
||||
// private FullTextSearchIndexer ftsIndexer;
|
||||
// /** the component providing searches of the indexed nodes */
|
||||
// private SearchService searcher;
|
||||
// /** the component giving direct access to <b>node</b> instances */
|
||||
// private NodeService nodeService;
|
||||
// /** set this to run the index recovery component */
|
||||
// private boolean executeFullRecovery;
|
||||
// /** set this on to keep checking for new transactions and never stop */
|
||||
// private boolean runContinuously;
|
||||
// /** set the time to wait between checking indexes */
|
||||
// private long waitTime;
|
||||
// /** controls how the L2 cache is used */
|
||||
// private CacheMode l2CacheMode;
|
||||
//
|
||||
// /**
|
||||
// * @return Returns the ID of the current (or last) transaction processed
|
||||
// */
|
||||
// public static String getCurrentTransactionId()
|
||||
// {
|
||||
// return currentTxnId;
|
||||
// }
|
||||
//
|
||||
// public FullIndexRecoveryComponent()
|
||||
// {
|
||||
// this.killThread = false;
|
||||
// this.executeFullRecovery = false;
|
||||
// this.runContinuously = false;
|
||||
// this.waitTime = 1000L;
|
||||
// this.l2CacheMode = CacheMode.REFRESH;
|
||||
//
|
||||
// // ensure that we kill the thread when the VM is shutting down
|
||||
// Runnable shutdownRunnable = new Runnable()
|
||||
// {
|
||||
// public void run()
|
||||
// {
|
||||
// killThread = true;
|
||||
// };
|
||||
// };
|
||||
// Thread shutdownThread = new Thread(shutdownRunnable);
|
||||
// Runtime.getRuntime().addShutdownHook(shutdownThread);
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * @return Returns true if the component has already been started
|
||||
// */
|
||||
// public static boolean isStarted()
|
||||
// {
|
||||
// return started;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * @param transactionService provide transactions to index each missed transaction
|
||||
// */
|
||||
// public void setTransactionService(TransactionService transactionService)
|
||||
// {
|
||||
// this.transactionService = transactionService;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * @param indexer the indexer that will be index
|
||||
// */
|
||||
// public void setIndexer(Indexer indexer)
|
||||
// {
|
||||
// this.indexer = indexer;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * @param ftsIndexer the FTS background indexer
|
||||
// */
|
||||
// public void setFtsIndexer(FullTextSearchIndexer ftsIndexer)
|
||||
// {
|
||||
// this.ftsIndexer = ftsIndexer;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * @param searcher component providing index searches
|
||||
// */
|
||||
// public void setSearcher(SearchService searcher)
|
||||
// {
|
||||
// this.searcher = searcher;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * @param nodeService provides information about nodes for indexing
|
||||
// */
|
||||
// public void setNodeService(NodeService nodeService)
|
||||
// {
|
||||
// this.nodeService = nodeService;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * Set this to <code>true</code> to initiate the full index recovery.
|
||||
// * <p>
|
||||
// * This used to default to <code>true</code> but is now false. Set this
|
||||
// * if the potentially long-running process of checking and fixing the
|
||||
// * indexes must be started.
|
||||
// *
|
||||
// * @param executeFullRecovery
|
||||
// */
|
||||
// public void setExecuteFullRecovery(boolean executeFullRecovery)
|
||||
// {
|
||||
// this.executeFullRecovery = executeFullRecovery;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * Set this to ensure that the process continuously checks for new transactions.
|
||||
// * If not, it will permanently terminate once it catches up with the current
|
||||
// * transactions.
|
||||
// *
|
||||
// * @param runContinuously true to never cease looking for new transactions
|
||||
// */
|
||||
// public void setRunContinuously(boolean runContinuously)
|
||||
// {
|
||||
// this.runContinuously = runContinuously;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * Set the time to wait between checking for new transaction changes in the database.
|
||||
// *
|
||||
// * @param waitTime the time to wait in milliseconds
|
||||
// */
|
||||
// public void setWaitTime(long waitTime)
|
||||
// {
|
||||
// this.waitTime = waitTime;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * Set the hibernate cache mode by name
|
||||
// *
|
||||
// * @see org.hibernate.CacheMode
|
||||
// */
|
||||
// public void setL2CacheMode(String l2CacheModeStr)
|
||||
// {
|
||||
// if (l2CacheModeStr.equals("GET"))
|
||||
// {
|
||||
// l2CacheMode = CacheMode.GET;
|
||||
// }
|
||||
// else if (l2CacheModeStr.equals("IGNORE"))
|
||||
// {
|
||||
// l2CacheMode = CacheMode.IGNORE;
|
||||
// }
|
||||
// else if (l2CacheModeStr.equals("NORMAL"))
|
||||
// {
|
||||
// l2CacheMode = CacheMode.NORMAL;
|
||||
// }
|
||||
// else if (l2CacheModeStr.equals("PUT"))
|
||||
// {
|
||||
// l2CacheMode = CacheMode.PUT;
|
||||
// }
|
||||
// else if (l2CacheModeStr.equals("REFRESH"))
|
||||
// {
|
||||
// l2CacheMode = CacheMode.REFRESH;
|
||||
// }
|
||||
// else
|
||||
// {
|
||||
// throw new IllegalArgumentException("Unrecognised Hibernate L2 cache mode: " + l2CacheModeStr);
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * Ensure that the index is up to date with the current state of the persistence layer.
|
||||
// * The full list of unique transaction change IDs is retrieved and used to detect
|
||||
// * which are not present in the index. All the node changes and deletions for the
|
||||
// * remaining transactions are then indexed.
|
||||
// */
|
||||
// public synchronized void reindex()
|
||||
// {
|
||||
// if (FullIndexRecoveryComponent.started)
|
||||
// {
|
||||
// throw new AlfrescoRuntimeException
|
||||
// ("Only one FullIndexRecoveryComponent may be used per VM and it may only be called once");
|
||||
// }
|
||||
//
|
||||
// // ensure that we don't redo this work
|
||||
// FullIndexRecoveryComponent.started = true;
|
||||
//
|
||||
// // work to mark the stores for full text reindexing
|
||||
// TransactionWork<Object> ftsReindexWork = new TransactionWork<Object>()
|
||||
// {
|
||||
// public Object doWork()
|
||||
// {
|
||||
// List<StoreRef> storeRefs = nodeService.getStores();
|
||||
// // reindex each store
|
||||
// for (StoreRef storeRef : storeRefs)
|
||||
// {
|
||||
// // check if the store exists
|
||||
// if (!nodeService.exists(storeRef))
|
||||
// {
|
||||
// // store does not exist
|
||||
// if (logger.isDebugEnabled())
|
||||
// {
|
||||
// logger.debug("Skipping reindex of non-existent store: " + storeRef);
|
||||
// }
|
||||
// continue;
|
||||
// }
|
||||
//
|
||||
// // prompt FTS to reindex the store
|
||||
// ftsIndexer.requiresIndex(storeRef);
|
||||
// }
|
||||
// // done
|
||||
// if (logger.isDebugEnabled())
|
||||
// {
|
||||
// logger.debug("Prompted FTS index on stores: " + storeRefs);
|
||||
// }
|
||||
// return null;
|
||||
// }
|
||||
// };
|
||||
// TransactionUtil.executeInNonPropagatingUserTransaction(transactionService, ftsReindexWork);
|
||||
//
|
||||
// // start full index recovery, if necessary
|
||||
// if (!this.executeFullRecovery)
|
||||
// {
|
||||
// if (logger.isDebugEnabled())
|
||||
// {
|
||||
// logger.debug("Full index recovery is off - quitting");
|
||||
// }
|
||||
// }
|
||||
// else
|
||||
// {
|
||||
// // set the state of the reindex
|
||||
// FullIndexRecoveryComponent.currentTxnId = START_TXN_ID;
|
||||
//
|
||||
// // start a stateful thread that will begin processing the reindexing the transactions
|
||||
// Runnable runnable = new ReindexRunner();
|
||||
// Thread reindexThread = new Thread(runnable);
|
||||
// // make it a daemon thread
|
||||
// reindexThread.setDaemon(true);
|
||||
// // it should not be a high priority
|
||||
// reindexThread.setPriority(Thread.MIN_PRIORITY);
|
||||
// // start it
|
||||
// reindexThread.start();
|
||||
//
|
||||
// if (logger.isDebugEnabled())
|
||||
// {
|
||||
// logger.debug("Full index recovery thread started: \n" +
|
||||
// " continuous: " + runContinuously);
|
||||
// }
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * Stateful thread runnable that executes reindex calls.
|
||||
// *
|
||||
// * @see FullIndexRecoveryComponent#reindexNodes()
|
||||
// *
|
||||
// * @author Derek Hulley
|
||||
// */
|
||||
// private class ReindexRunner implements Runnable
|
||||
// {
|
||||
// public void run()
|
||||
// {
|
||||
// // keep this thread going permanently
|
||||
// while (!killThread)
|
||||
// {
|
||||
// try
|
||||
// {
|
||||
// // reindex nodes
|
||||
// List<String> txnsIndexed = FullIndexRecoveryComponent.this.reindexNodes();
|
||||
// // reindex missing content
|
||||
// @SuppressWarnings("unused")
|
||||
// int missingContentCount = FullIndexRecoveryComponent.this.reindexMissingContent();
|
||||
// // check if the process should terminate
|
||||
// if (txnsIndexed.size() == 0 && !runContinuously)
|
||||
// {
|
||||
// // the thread has caught up with all the available work and should not
|
||||
// // run continuously
|
||||
// if (logger.isDebugEnabled())
|
||||
// {
|
||||
// logger.debug("Thread quitting - no more available indexing to do: \n" +
|
||||
// " last txn: " + FullIndexRecoveryComponent.getCurrentTransactionId());
|
||||
// }
|
||||
// break;
|
||||
// }
|
||||
// // brief pause
|
||||
// synchronized(FullIndexRecoveryComponent.this)
|
||||
// {
|
||||
// FullIndexRecoveryComponent.this.wait(waitTime);
|
||||
// }
|
||||
// }
|
||||
// catch (InterruptedException e)
|
||||
// {
|
||||
// // ignore
|
||||
// }
|
||||
// catch (Throwable e)
|
||||
// {
|
||||
// if (killThread)
|
||||
// {
|
||||
// // the shutdown may have caused the exception - ignore it
|
||||
// }
|
||||
// else
|
||||
// {
|
||||
// // we are still a go; report it
|
||||
// logger.error("Reindex failure", e);
|
||||
// }
|
||||
// }
|
||||
// }
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * @return Returns the number of documents reindexed
|
||||
// */
|
||||
// private int reindexMissingContent()
|
||||
// {
|
||||
// int count = 0;
|
||||
// for (StoreRef storeRef : storeRefs)
|
||||
// {
|
||||
// count += reindexMissingContent(storeRef);
|
||||
// }
|
||||
// return count;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * @param storeRef the store to check for missing content
|
||||
// * @return Returns the number of documents reindexed
|
||||
// */
|
||||
// private int reindexMissingContent(StoreRef storeRef)
|
||||
// {
|
||||
// SearchParameters sp = new SearchParameters();
|
||||
// sp.addStore(storeRef);
|
||||
//
|
||||
// // search for it in the index
|
||||
// String query = "TEXT:" + LuceneIndexerImpl.NOT_INDEXED_CONTENT_MISSING;
|
||||
// sp.setLanguage(SearchService.LANGUAGE_LUCENE);
|
||||
// sp.setQuery(query);
|
||||
// ResultSet results = null;
|
||||
// try
|
||||
// {
|
||||
// results = searcher.query(sp);
|
||||
//
|
||||
// int count = 0;
|
||||
// // loop over the results and get the details of the nodes that have missing content
|
||||
// List<ChildAssociationRef> assocRefs = results.getChildAssocRefs();
|
||||
// for (ChildAssociationRef assocRef : assocRefs)
|
||||
// {
|
||||
// final NodeRef childNodeRef = assocRef.getChildRef();
|
||||
// // prompt for a reindex - it might fail again, but we just keep plugging away
|
||||
// TransactionWork<Object> reindexWork = new TransactionWork<Object>()
|
||||
// {
|
||||
// public Object doWork()
|
||||
// {
|
||||
// indexer.updateNode(childNodeRef);
|
||||
// return null;
|
||||
// }
|
||||
// };
|
||||
// TransactionUtil.executeInNonPropagatingUserTransaction(transactionService, reindexWork);
|
||||
// count++;
|
||||
// }
|
||||
// // done
|
||||
// if (logger.isDebugEnabled())
|
||||
// {
|
||||
// logger.debug("Reindexed missing content: \n" +
|
||||
// " store: " + storeRef + "\n" +
|
||||
// " node count: " + count);
|
||||
// }
|
||||
// return count;
|
||||
// }
|
||||
// finally
|
||||
// {
|
||||
// if (results != null)
|
||||
// {
|
||||
// results.close();
|
||||
// }
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * @return Returns the transaction ID just reindexed, i.e. where some work was performed
|
||||
// */
|
||||
// private List<String> reindexNodes()
|
||||
// {
|
||||
// // get a list of all transactions still requiring a check
|
||||
// List<String> txnsToCheck = getNextChangeTxnIds(FullIndexRecoveryComponent.currentTxnId);
|
||||
//
|
||||
// // loop over each transaction
|
||||
// for (String changeTxnId : txnsToCheck)
|
||||
// {
|
||||
// reindexNodes(changeTxnId);
|
||||
// }
|
||||
//
|
||||
// // done
|
||||
// return txnsToCheck;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * Reindexes changes specific to the change transaction ID.
|
||||
// * <p>
|
||||
// * <b>All exceptions are absorbed.</b>
|
||||
// */
|
||||
// private void reindexNodes(final String changeTxnId)
|
||||
// {
|
||||
// /*
|
||||
// * This must execute each within its own transaction.
|
||||
// * The cache size is therefore not an issue.
|
||||
// */
|
||||
// TransactionWork<Object> reindexWork = new TransactionWork<Object>()
|
||||
// {
|
||||
// public Object doWork() throws Exception
|
||||
// {
|
||||
// // perform the work in a Hibernate callback
|
||||
// HibernateCallback callback = new ReindexCallback(changeTxnId);
|
||||
// getHibernateTemplate().execute(callback);
|
||||
// // done
|
||||
// return null;
|
||||
// }
|
||||
// };
|
||||
// try
|
||||
// {
|
||||
// TransactionUtil.executeInNonPropagatingUserTransaction(transactionService, reindexWork);
|
||||
// }
|
||||
// catch (Throwable e)
|
||||
// {
|
||||
// logger.error("Transaction reindex failed: \n" +
|
||||
// " txn: " + changeTxnId,
|
||||
// e);
|
||||
// }
|
||||
// finally
|
||||
// {
|
||||
// // Up the current transaction now, in case the process fails at this point.
|
||||
// // This will prevent the transaction from being processed again.
|
||||
// // This applies to failures as well, which should be dealt with externally
|
||||
// // and having the entire process start again, e.g. such as a system reboot
|
||||
// currentTxnId = changeTxnId;
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * Stateful inner class that implements a single reindex call for a given store
|
||||
// * and transaction.
|
||||
// * <p>
|
||||
// * It must be called within its own transaction.
|
||||
// *
|
||||
// * @author Derek Hulley
|
||||
// */
|
||||
// private class ReindexCallback implements HibernateCallback
|
||||
// {
|
||||
// private final String changeTxnId;
|
||||
//
|
||||
// public ReindexCallback(String changeTxnId)
|
||||
// {
|
||||
// this.changeTxnId = changeTxnId;
|
||||
// }
|
||||
//
|
||||
// /**
|
||||
// * Changes the L2 cache usage before reindexing for each store
|
||||
// *
|
||||
// * @see #reindexNodes(StoreRef, String)
|
||||
// */
|
||||
// public Object doInHibernate(Session session)
|
||||
// {
|
||||
// // set the way the L2 cache is used
|
||||
// getSession().setCacheMode(l2CacheMode);
|
||||
//
|
||||
// // reindex each store
|
||||
// for (StoreRef storeRef : storeRefs)
|
||||
// {
|
||||
// if (!nodeService.exists(storeRef))
|
||||
// {
|
||||
// // the store is not present
|
||||
// continue;
|
||||
// }
|
||||
// // reindex for store
|
||||
// reindexNodes(storeRef, changeTxnId);
|
||||
// }
|
||||
// // done
|
||||
// return null;
|
||||
// }
|
||||
//
|
||||
// private void reindexNodes(StoreRef storeRef, String changeTxnId)
|
||||
// {
|
||||
// // check if we need to perform this operation
|
||||
// SearchParameters sp = new SearchParameters();
|
||||
// sp.addStore(storeRef);
|
||||
//
|
||||
// // search for it in the index
|
||||
// String query = "TX:\"" + changeTxnId + "\"";
|
||||
// sp.setLanguage(SearchService.LANGUAGE_LUCENE);
|
||||
// sp.setQuery(query);
|
||||
// ResultSet results = null;
|
||||
// try
|
||||
// {
|
||||
// results = searcher.query(sp);
|
||||
// // did the index have any of these changes?
|
||||
// if (results.length() > 0)
|
||||
// {
|
||||
// // the transaction has an entry in the index - assume that it was
|
||||
// // atomically correct
|
||||
// if (logger.isDebugEnabled())
|
||||
// {
|
||||
// logger.debug("Transaction present in index - no indexing required: \n" +
|
||||
// " store: " + storeRef + "\n" +
|
||||
// " txn: " + changeTxnId);
|
||||
// }
|
||||
// return;
|
||||
// }
|
||||
// }
|
||||
// finally
|
||||
// {
|
||||
// if (results != null)
|
||||
// {
|
||||
// results.close();
|
||||
// }
|
||||
// }
|
||||
// // the index has no record of this
|
||||
// // were there any changes, or is it all just deletions?
|
||||
// int changedCount = getChangedNodeStatusesCount(storeRef, changeTxnId);
|
||||
// if (changedCount == 0)
|
||||
// {
|
||||
// // no nodes were changed in the transaction, i.e. they are only deletions
|
||||
// // the index is quite right not to have any entries for the transaction
|
||||
// if (logger.isDebugEnabled())
|
||||
// {
|
||||
// logger.debug("Transaction only has deletions - no indexing required: \n" +
|
||||
// " store: " + storeRef + "\n" +
|
||||
// " txn: " + changeTxnId);
|
||||
// }
|
||||
// return;
|
||||
// }
|
||||
//
|
||||
// // process the deletions relevant to the txn and the store
|
||||
// List<NodeStatus> deletedNodeStatuses = getDeletedNodeStatuses(storeRef, changeTxnId);
|
||||
// for (NodeStatus status : deletedNodeStatuses)
|
||||
// {
|
||||
// NodeRef nodeRef = new NodeRef(storeRef, status.getKey().getGuid());
|
||||
// // only the child node ref is relevant
|
||||
// ChildAssociationRef assocRef = new ChildAssociationRef(
|
||||
// ContentModel.ASSOC_CHILDREN,
|
||||
// null,
|
||||
// null,
|
||||
// nodeRef);
|
||||
// indexer.deleteNode(assocRef);
|
||||
// }
|
||||
//
|
||||
// // process additions
|
||||
// List<NodeStatus> changedNodeStatuses = getChangedNodeStatuses(storeRef, changeTxnId);
|
||||
// for (NodeStatus status : changedNodeStatuses)
|
||||
// {
|
||||
// NodeRef nodeRef = new NodeRef(storeRef, status.getKey().getGuid());
|
||||
// // get the primary assoc for the node
|
||||
// ChildAssociationRef primaryAssocRef = nodeService.getPrimaryParent(nodeRef);
|
||||
// // reindex
|
||||
// indexer.createNode(primaryAssocRef);
|
||||
// }
|
||||
//
|
||||
// // done
|
||||
// if (logger.isDebugEnabled())
|
||||
// {
|
||||
// logger.debug("Transaction reindexed: \n" +
|
||||
// " store: " + storeRef + "\n" +
|
||||
// " txn: " + changeTxnId + "\n" +
|
||||
// " deletions: " + deletedNodeStatuses.size() + "\n" +
|
||||
// " modifications: " + changedNodeStatuses.size());
|
||||
// }
|
||||
// }
|
||||
// };
|
||||
//
|
||||
// /**
|
||||
// * Retrieve all transaction IDs that are greater than the given transaction ID.
|
||||
// *
|
||||
// * @param currentTxnId the transaction ID that must be less than all returned results
|
||||
// * @return Returns an ordered list of transaction IDs
|
||||
// */
|
||||
// @SuppressWarnings("unchecked")
|
||||
// public List<String> getNextChangeTxnIds(final String currentTxnId)
|
||||
// {
|
||||
// HibernateCallback callback = new HibernateCallback()
|
||||
// {
|
||||
// public Object doInHibernate(Session session)
|
||||
// {
|
||||
// Query query = session.getNamedQuery(QUERY_GET_NEXT_CHANGE_TXN_IDS);
|
||||
// query.setString("currentTxnId", currentTxnId)
|
||||
// .setReadOnly(true);
|
||||
// return query.list();
|
||||
// }
|
||||
// };
|
||||
// List<String> queryResults = (List<String>) getHibernateTemplate().execute(callback);
|
||||
// // done
|
||||
// return queryResults;
|
||||
// }
|
||||
//
|
||||
// @SuppressWarnings("unchecked")
|
||||
// public int getChangedNodeStatusesCount(final StoreRef storeRef, final String changeTxnId)
|
||||
// {
|
||||
// HibernateCallback callback = new HibernateCallback()
|
||||
// {
|
||||
// public Object doInHibernate(Session session)
|
||||
// {
|
||||
// Query query = session.getNamedQuery(QUERY_GET_CHANGED_NODE_STATUSES_COUNT);
|
||||
// query.setString("storeProtocol", storeRef.getProtocol())
|
||||
// .setString("storeIdentifier", storeRef.getIdentifier())
|
||||
// .setString("changeTxnId", changeTxnId)
|
||||
// .setReadOnly(true);
|
||||
// return query.uniqueResult();
|
||||
// }
|
||||
// };
|
||||
// Integer changeCount = (Integer) getHibernateTemplate().execute(callback);
|
||||
// // done
|
||||
// return changeCount.intValue();
|
||||
// }
|
||||
//
|
||||
// @SuppressWarnings("unchecked")
|
||||
// public List<NodeStatus> getChangedNodeStatuses(final StoreRef storeRef, final String changeTxnId)
|
||||
// {
|
||||
// HibernateCallback callback = new HibernateCallback()
|
||||
// {
|
||||
// public Object doInHibernate(Session session)
|
||||
// {
|
||||
// Query query = session.getNamedQuery(QUERY_GET_CHANGED_NODE_STATUSES);
|
||||
// query.setString("storeProtocol", storeRef.getProtocol())
|
||||
// .setString("storeIdentifier", storeRef.getIdentifier())
|
||||
// .setString("changeTxnId", changeTxnId)
|
||||
// .setReadOnly(true);
|
||||
// return query.list();
|
||||
// }
|
||||
// };
|
||||
// List<NodeStatus> queryResults = (List) getHibernateTemplate().execute(callback);
|
||||
// // done
|
||||
// return queryResults;
|
||||
// }
|
||||
//
|
||||
// @SuppressWarnings("unchecked")
|
||||
// public List<NodeStatus> getDeletedNodeStatuses(final StoreRef storeRef, final String changeTxnId)
|
||||
// {
|
||||
// HibernateCallback callback = new HibernateCallback()
|
||||
// {
|
||||
// public Object doInHibernate(Session session)
|
||||
// {
|
||||
// Query query = session.getNamedQuery(QUERY_GET_DELETED_NODE_STATUSES);
|
||||
// query.setString("storeProtocol", storeRef.getProtocol())
|
||||
// .setString("storeIdentifier", storeRef.getIdentifier())
|
||||
// .setString("changeTxnId", changeTxnId)
|
||||
// .setReadOnly(true);
|
||||
// return query.list();
|
||||
// }
|
||||
// };
|
||||
// List<NodeStatus> queryResults = (List) getHibernateTemplate().execute(callback);
|
||||
// // done
|
||||
// return queryResults;
|
||||
// }
|
||||
//}
|
@@ -307,23 +307,6 @@ public class IntegrityTest extends TestCase
|
||||
logger.error("Method commented out: testRemoveSourcesOfMandatoryAssocs");
|
||||
}
|
||||
|
||||
public void testDuplicateTargetAssocs() throws Exception
|
||||
{
|
||||
NodeRef parent = createNode("source", TEST_TYPE_WITH_CHILD_ASSOCS, null);
|
||||
NodeRef child1 = createNode("child1", TEST_TYPE_WITHOUT_ANYTHING, null);
|
||||
NodeRef child2 = createNode("child2", TEST_TYPE_WITHOUT_ANYTHING, null);
|
||||
NodeRef child3 = createNode("child3", TEST_TYPE_WITHOUT_ANYTHING, null);
|
||||
|
||||
// satisfy the one-to-one
|
||||
nodeService.addChild(parent, child3, TEST_ASSOC_CHILD_ONE_ONE, QName.createQName(NAMESPACE, "mandatoryChild"));
|
||||
|
||||
// create the non-duplicate assocs
|
||||
nodeService.addChild(parent, child1, TEST_ASSOC_CHILD_ZEROMANY_ZEROMANY, QName.createQName(NAMESPACE, "dupli_cate"));
|
||||
nodeService.addChild(parent, child2, TEST_ASSOC_CHILD_ZEROMANY_ZEROMANY, QName.createQName(NAMESPACE, "dupli_cate"));
|
||||
|
||||
checkIntegrityExpectFailure("Failed to detect duplicate association names", 1);
|
||||
}
|
||||
|
||||
public void testCreateSourceOfAssocsWithMandatoryTargetsPresent() throws Exception
|
||||
{
|
||||
NodeRef source = createNode("abc", TEST_TYPE_WITH_ASSOCS, null);
|
||||
|
Reference in New Issue
Block a user