Dave Ward a2580451b9 Merged V3.3-BUG-FIX to HEAD
21242: ALF-2879: XAM Connector changes
      - Fixed setting of store name
      - Added properties:
         - xam.archive.retentionPeriodDays=0
         - xam.archive.addLock=true
   21244: ALF-2879: Updated readme.txt
   21262: ALF-3611 - tweak AVM orphan reaper test (PurgeTestP)
   21277: Fix ALF-889 - WCM/AVM folder disappears when cut-and-paste into itself
      - fixed cycle check before rename/move
      - added -ve unit test
      - externalized existing cycle error messages
   21284: ALF-2879: XAM Connector changes
      - Listen to store selector policies (incl. workaround for policy listening)
      - Set XAM retention (code is setting the value, but not successfully tested against test servers)
   21285: StoreSelectorPolicies.OnContentMovedPolicy is deprecated. Use StoreSelectorPolicies.AfterMoveContentPolicy.
      - Deprecated so old policy still exists and works
      - Will remove for 3.4 (maybe)
   21293: Fix ALF-3245: stream not closed in DictionaryBootstrap.onDictionaryInit()
   21303: ALF-2879: XAM Connector changes
      - Throw ContentIOException if setBaseRetention fails
   21313: ALF-2879: XAM Connector changes
      - Round ms to nearest second
      - Log actual ms value being set for retention
   21322: Fix AVMNodeService.createNode to close output stream and avoid "Too many open files" (also add example AVMFileFolderPerformanceTester)
   21331: ALF-2879: XAM Connector changes
      - Removed duplicate setting of base retention
      - Left code hooks for setting of other metadata
   21368: Merged V3.3 to V3.3-BUG-FIX
      21213: Merged DEV/TEMPORARY to V3.3
         21200: ALF-2978: IMAP cannot bind to all the interfaces (0.0.0.0)
            “imap.server.host” property can be used for setting IP address / network adapter to listen on for IMAP protocol.
      21219: Merged PATCHES/V3.2.1 to V3.3
         21216: ALF-3779: A few bug fixes to --FOREACH handling in SchemaBootstrap
            - New system.upgrade.default.batchsize property to control overall default batch size
            - Added in a few more missed --FOREACH markers
         21211: ALF-3779: Remaining scripts converted to use --FOREACH (as logs finally provided by test prove that they need it too!)
         21210: (RECORD ONLY) Incremented version label
         21209: ALF-3779: Solution to allow batching of mass updates in upgrade scripts into smaller transactions
            - A special preceding comment in this format specifies a numeric table column to control the batching and a global property specifying the batch size
               --FOREACH table.column batch.size.property
            - If the property isn't specified in alfresco-global.properties, the default batch size is 10,000
            - INSERT / UPDATE / DELETE statements can then tack on extra conditions on ${LOWERBOUND} and ${UPPERBOUND} variables. E.g.
               WHERE n.id >= ${LOWERBOUND} AND n.id <= ${UPPERBOUND}
            - The statements are substituted and executed for each batch range up to the maximum value of the column
            - 2.1 and 2.2 MySQL upgrades reimplemented this way
         21207: Extra debug logging to track index triggering activity
      21295: Merged HEAD to V3.3
         21255: Parameter encoding
      21298: Merged V3.2 to V3.3
         21297: ALF-3889: JBPMDeployProcessServlet is now disabled by default and enabled with this in alfresco-global.properties
            system.workflow.deployservlet.enabled=true
      21317: dod5015: Parameter encoding
      21363: Merged PATCHES/V3.2.1 to V3.3
         21338: (RECORD ONLY) Incremented version label
         21335: ALF-3779: Correction to error in --FOREACH range restriction for UPDATE statement
         21290: ALF-3960: ArrayIndexOutOfBoundsException when we set mergerMergeFactor > mergerTargetOverlays
         21278: (RECORD ONLY) Merged PATCHES/V3.1.2 to PATCHES/V3.2.1
            21264: ALF-3889: JBPMDeployProcessServlet not accessible by default
               - Should only be enabled in development environment
      21364: Merged PATCHES/V3.1.2 to V3.3 (RECORD ONLY)
         21264: ALF-3889: JBPMDeployProcessServlet not accessible by default
            - Should only be enabled in development environment
      21365: Merged PATCHES/V3.2.0 to V3.3 (RECORD ONLY)
         21276: Merged PATCHES/V3.1.2 to PATCHES/V3.2.0
            21264: ALF-3889: JBPMDeployProcessServlet not accessible by default
               - Should only be enabled in development environment
      21366: Merged PATCHES/V3.2.r to V3.3 (RECORD ONLY)
         21279: Merged PATCHES/V3.1.2 to PATCHES/V3.2.r
            21264: ALF-3889: JBPMDeployProcessServlet not accessible by default
               - Should only be enabled in development environment
      21367: Merged PATCHES/V3.3.1 to V3.3 (RECORD ONLY)
         21343: Incremented version label
         21342: ALF-3997: Merged V3.3-BUG-FIX to PATCHES/V3.3.1
            20623: Fix for ALF-3188 : Access Denied when updating doc via CIFS
         21282: Merged PATCHES/V3.1.2 to PATCHES/V3.3.1
            21264: ALF-3889: JBPMDeployProcessServlet not accessible by default
               - Should only be enabled in development environment
         21239: Created hotfix branch off ENTERPRISE/V3.3.1


git-svn-id: https://svn.alfresco.com/repos/alfresco-enterprise/alfresco/HEAD/root@21369 c4b6b30b-aa2e-2d43-bbcb-ca4b014f7261
2010-07-22 18:20:24 +00:00

442 lines
15 KiB
Java

/*
* Copyright (C) 2005-2010 Alfresco Software Limited.
*
* This file is part of Alfresco
*
* Alfresco is free software: you can redistribute it and/or modify
* it under the terms of the GNU Lesser General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* Alfresco is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public License
* along with Alfresco. If not, see <http://www.gnu.org/licenses/>. */
package org.alfresco.repo.avm;
import java.util.LinkedList;
import java.util.List;
import org.alfresco.repo.domain.avm.AVMHistoryLinkEntity;
import org.alfresco.repo.domain.avm.AVMMergeLinkEntity;
import org.alfresco.repo.domain.permissions.Acl;
import org.alfresco.repo.transaction.RetryingTransactionHelper.RetryingTransactionCallback;
import org.alfresco.service.cmr.repository.ContentData;
import org.alfresco.service.transaction.TransactionService;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
/**
* This is the background thread for reaping no longer referenced nodes in the AVM repository. These orphans arise from
* purge operations.
*
* @author britt
*/
public class OrphanReaper
{
public void execute()
{
synchronized (this)
{
if (fRunning)
{
if (fgLogger.isDebugEnabled())
{
fgLogger.debug("OrphanReaper is already running - just return");
}
return;
}
fRunning = true;
if (fgLogger.isTraceEnabled())
{
fgLogger.trace("Start running OrphanReaper ...");
}
}
try
{
do
{
doBatch();
if (fDone)
{
if (fgLogger.isTraceEnabled())
{
fgLogger.trace("OrphanReaper is done - just return");
}
return;
}
try
{
if (fgLogger.isTraceEnabled())
{
fgLogger.trace("OrphanReaper is not done - sleep for "+fActiveBaseSleep+" ms");
}
Thread.sleep(fActiveBaseSleep);
}
catch (InterruptedException e)
{
fgLogger.warn("OrphanReaper was interrupted - do nothing: "+e);
// Do nothing.
}
}
while (fActive);
}
finally
{
synchronized (this)
{
fRunning = false;
if (fgLogger.isTraceEnabled())
{
fgLogger.trace("... finish running OrphanReaper");
}
}
}
}
private Log fgLogger = LogFactory.getLog(OrphanReaper.class);
/**
* The Transaction Service
*/
private TransactionService fTransactionService;
/**
* Active base sleep interval.
*/
private long fActiveBaseSleep;
/**
* Batch size.
*/
private int fBatchSize;
/**
* Whether we are currently active, ie have work queued up.
*/
private boolean fActive;
/**
* The maximum length of the queue.
*/
private int fQueueLength;
/**
* The linked list containing ids of nodes that are purgable.
*/
private LinkedList<Long> fPurgeQueue;
private boolean fDone = false;
private boolean fRunning = false;
/**
* Create one with default parameters.
*/
public OrphanReaper()
{
fActiveBaseSleep = 1000;
fBatchSize = 50;
fQueueLength = 1000;
fActive = false;
}
// Setters for configuration.
/**
* Set the active base sleep interval.
*
* @param interval
* The interval to set in ms.
*/
public void setActiveBaseSleep(long interval)
{
fActiveBaseSleep = interval;
}
/**
* Set the batch size.
*
* @param size
* The batch size to set.
*/
public void setBatchSize(int size)
{
fBatchSize = size;
}
/**
* Set the transaction service.
*
* @param transactionService
* The service.
*/
public void setTransactionService(TransactionService transactionService)
{
fTransactionService = transactionService;
}
/**
* Set the maximum size of the queue of purgeable nodes.
*
* @param queueLength
* The max length.
*/
public void setMaxQueueLength(int queueLength)
{
fQueueLength = queueLength;
}
/**
* Start things up after configuration is complete.
*/
// public void init()
// {
// fThread = new Thread(this);
// fThread.start();
// }
/**
* Shutdown the reaper. This needs to be called when the application shuts down.
*/
public void shutDown()
{
fDone = true;
}
/**
* Sit in a loop, periodically querying for orphans. When orphans are found, unhook them in bite sized batches.
*/
// public void run()
// {
// while (!fDone)
// {
// synchronized (this)
// {
// try
// {
// wait(fActive? fActiveBaseSleep : fInactiveBaseSleep);
// }
// catch (InterruptedException ie)
// {
// // Do nothing.
// }
// doBatch();
// }
// }
// }
/**
* This is really for debugging and testing. Allows another thread to mark the orphan reaper busy so that it can
* monitor for it's being done.
*/
public void activate()
{
fActive = true;
}
/**
* See if the reaper is actively reaping.
*
* @return Whether this is actively reaping.
*/
public boolean isActive()
{
return fActive;
}
/**
* Do a batch of cleanup work.
*/
public void doBatch()
{
class TxnWork implements RetryingTransactionCallback<Object>
{
public Object execute() throws Exception
{
if (fgLogger.isTraceEnabled())
{
fgLogger.trace("Orphan reaper doBatch: batchSize="+fBatchSize+", maxQueueLength="+fQueueLength+", fActiveBaseSleep="+fActiveBaseSleep);
}
if (fPurgeQueue == null)
{
List<AVMNode> nodes = AVMDAOs.Instance().fAVMNodeDAO.getOrphans(fQueueLength);
if (nodes.size() == 0)
{
if (fgLogger.isTraceEnabled())
{
fgLogger.trace("Nothing to purge (set fActive = false)");
}
fActive = false;
return null;
}
fPurgeQueue = new LinkedList<Long>();
for (AVMNode node : nodes)
{
fPurgeQueue.add(node.getId());
}
if (fgLogger.isDebugEnabled())
{
fgLogger.debug("Queue was empty so got more orphans from DB. Orphan queue size = "+fPurgeQueue.size());
}
}
else
{
if (fgLogger.isDebugEnabled())
{
fgLogger.debug("Queue was not empty. Orphan queue size = "+fPurgeQueue.size());
}
}
fActive = true;
int reapCnt = 0;
long start = System.currentTimeMillis();
for (int i = 0; i < fBatchSize; i++)
{
if (fPurgeQueue.size() == 0)
{
if (fgLogger.isTraceEnabled())
{
fgLogger.trace("Purge queue is empty (fpurgeQueue size = "+fPurgeQueue.size()+")");
}
fPurgeQueue = null;
break;
}
Long nodeId = fPurgeQueue.removeFirst();
AVMNode node = AVMDAOs.Instance().fAVMNodeDAO.getByID(nodeId);
if (node == null)
{
// eg. cluster, multiple reapers
fgLogger.warn("Node ["+nodeId+"] not found - assume multiple reapers ...");
continue;
}
// Save away the ancestor and merged from fields from this node.
AVMNode ancestor = null;
AVMHistoryLinkEntity hlEntity = AVMDAOs.Instance().newAVMNodeLinksDAO.getHistoryLinkByDescendent(node.getId());
if (hlEntity != null)
{
ancestor = AVMDAOs.Instance().fAVMNodeDAO.getByID(hlEntity.getAncestorNodeId());
AVMDAOs.Instance().newAVMNodeLinksDAO.deleteHistoryLink(hlEntity.getAncestorNodeId(), hlEntity.getDescendentNodeId());
}
AVMNode mergedFrom = null;
AVMMergeLinkEntity mlEntity = AVMDAOs.Instance().newAVMNodeLinksDAO.getMergeLinkByTo(node.getId());
if (mlEntity != null)
{
mergedFrom = AVMDAOs.Instance().fAVMNodeDAO.getByID(mlEntity.getMergeFromNodeId());
AVMDAOs.Instance().newAVMNodeLinksDAO.deleteMergeLink(mlEntity.getMergeFromNodeId(), mlEntity.getMergeToNodeId());
}
// Get all the nodes that have this node as ancestor.
List<AVMHistoryLinkEntity> hlEntities = AVMDAOs.Instance().newAVMNodeLinksDAO.getHistoryLinksByAncestor(node.getId());
for (AVMHistoryLinkEntity link : hlEntities)
{
AVMNode desc = AVMDAOs.Instance().fAVMNodeDAO.getByID(link.getDescendentNodeId());
if (desc != null)
{
desc.setAncestor(ancestor);
if (desc.getMergedFrom() == null)
{
desc.setMergedFrom(mergedFrom);
}
}
AVMDAOs.Instance().newAVMNodeLinksDAO.deleteHistoryLink(link.getAncestorNodeId(), link.getDescendentNodeId());
}
// Get all the nodes that have this node as mergedFrom
List<AVMMergeLinkEntity> mlEntities = AVMDAOs.Instance().newAVMNodeLinksDAO.getMergeLinksByFrom(node.getId());
for (AVMMergeLinkEntity link : mlEntities)
{
AVMNode mto = AVMDAOs.Instance().fAVMNodeDAO.getByID(link.getMergeToNodeId());
if (mto != null)
{
mto.setMergedFrom(ancestor);
}
AVMDAOs.Instance().newAVMNodeLinksDAO.deleteMergeLink(link.getMergeFromNodeId(), link.getMergeToNodeId());
}
// Get rid of all properties belonging to this node.
AVMDAOs.Instance().fAVMNodeDAO.deleteProperties(node.getId());
// Get rid of all aspects belonging to this node.
AVMDAOs.Instance().fAVMNodeDAO.deleteAspects(node.getId());
// Get rid of ACL.
@SuppressWarnings("unused")
Acl acl = node.getAcl();
node.setAcl(null);
// Unused acls will be garbage collected
// Many acls will be shared
// Extra work for directories.
if (node.getType() == AVMNodeType.PLAIN_DIRECTORY || node.getType() == AVMNodeType.LAYERED_DIRECTORY)
{
// First get rid of all child entries for the node.
AVMDAOs.Instance().fChildEntryDAO.deleteByParent(node);
}
else if (node.getType() == AVMNodeType.PLAIN_FILE)
{
PlainFileNode file = (PlainFileNode)node;
if (file.isLegacyContentData())
{
// We quickly convert the old ContentData to the new storage
ContentData contentData = file.getContentData();
file.setContentData(contentData);
}
Long contentDataId = file.getContentDataId();
if (contentDataId != null)
{
// The ContentDataDAO will take care of dereferencing and cleanup
AVMDAOs.Instance().contentDataDAO.deleteContentData(contentDataId);
}
}
// Finally, delete it
AVMDAOs.Instance().fAVMNodeDAO.delete(node);
if (fgLogger.isTraceEnabled())
{
fgLogger.trace("Deleted Node ["+node.getId()+"]");
}
reapCnt++;
}
if (fgLogger.isDebugEnabled())
{
fgLogger.debug("Reaped "+reapCnt+" nodes in "+(System.currentTimeMillis()-start)+" msecs");
}
return null;
}
}
try
{
if (!fTransactionService.isReadOnly())
{
fTransactionService.getRetryingTransactionHelper().doInTransaction(new TxnWork());
}
}
catch (Exception e)
{
fgLogger.error("Garbage collector error", e);
}
}
}