Mark Rogers af1a735c98 Merge DM-DM_deployment to HEAD
17716 : Copied alfresco/HEAD to alfresco/BRANCHES/DEV/BRIAN/DM-DM_deployment.
  17721 : Moved alfresco/BRANCHES/DEV/BRIAN/DM-DM_deployment to alfresco/BRANCHES/DEV/DM-DM_deployment.
  17809 : Bare Bones of TransferService.
  17813 : more bones
  17824 : bootstrap of the transfer spaces
  17826 : Added username, path, and a getter for the password.
  17832 : createTransferTarget, getTransferTargets() working.
  17836 : transfer target 
    - duplicate name detection
    - delete transfer target by name
    - get transfer target by name
    - TransferException added along with transfer-service.properties
  17840 : transfer target
   - enable / disable
   - update properties
  17851 : Added the notion of a transmitter for the transfer service, and an HttpClient implementation of it.
  17852 : Added the web script needed for the receiving end of a transfer, along with a command processor for clients to 
check availability of the service and credentials.
  17856 : Added a Mockito-based test for HttpClientTransmitterImpl
  17857 : Corrected javadoc for test class
  17858 : Added test for overriding SSL socket factory
  17859 : Wired up the transfer service to the transfer transmitter.
  17896 : Rework to spring for app:transferDefinitions it's now app:transfer_definitio
    - Add throws clauses to TransferService
    - Bare bones of interface for begin and sendManifest
  17944 : Work in progress on manifest file
  17947 : added parent and child associations to the snapshot.
  17956 : Now has the parent path implemented and introduces the TransferManifestNodeFactory.
  17965 : Added content, MLText and collections.
  17978 : addition of source and target peer associations.
  17982 : Fixing parentNode information.
  18008 : XML Manifest Reader checkpoint. (Still incomplete but lots working)
  18040 : ParentPath is transmitted, fixes for source and target assocs.
  18048 : SAIL-30: Initial commit of functionality for begin, sendManifest, and sendContent parts of the transfer process
  18049 : Fix for parse of Locale type.
  18054 : Added TransferManifestNodeHelper and more tests
  18066 : Work in progress check in
   - Implemented the content chunker
 - sketched out more interfaces on TransferTransmitter.
 - Please note that the chunker is not yet connected to the manifest file, that will come next.
r18069 : Wired up manifest reader to content chunker.
r18089 : Fiest cut of callback interface for review.
r18091 : added hashCode implementation which was missing from ContentData
r18095 : Start of the server-side commit. Note that this is an interim commit - not tested.
r18096 : Initial entry of a ContentData implementation of HttpClient's "Part"
r18155 : Work in progress check in.
    TransferEvent - incomplete
    HttpClientTransmitter - first cut complete (not tested)
    Server side - first cut complete (not tested)
  18156 : TransferMessage missing from last check in.
  18166 : check in command processors
  18167 : Work primarily on the transfer commit operation
  18170 : corrected spring errors.
  18176 : Further testing and fixing of transfer commit
  18206 : Work in progress.
  18236 : Work in progress - generally adding debug logging and sorting out exception handlers.
  18240 : Fix to call "end" correctly after exception is thrown with "commit"
  18242 : Aligning the manifest part names.
  18243 : PostSnapshot calls the correct method on the receiver service.
  18267 : First node has transferred.
  18274 : Fixing abort to call end, debug statements, formatting code brackest
  18275 : First code to handle updates. Also improved error messages passed back to client
  18289 : Checked in work in progress.  Content upload not working.
  18290 : Update to ensure file type.
  18300 : Added more log output and some of the error messages.
  18301 : Work in progress
  18302 : Added log output
  18307 : Added a noddy transfer action
  18315 : Sprint 2 complete - transfer and update one node.
  18354 : Now the manifest file has deleted nodes.
 - Adding copyright headers
  18384 : Plumbing for unit tests on one box.
  18416 : First end to end unit test working.   (one node create and update)
  18421 : Added path based update test and many send test.
  18458 : Added the functionality to transfer deleted and restored nodes.
  18481 : Implementation of transferAsync
  18491 : SAIL-32, SAIL-35
   - Added node crawler for DM-DM Transfer F6 and F9.
  18620 : Basic transfer report implementation

git-svn-id: https://svn.alfresco.com/repos/alfresco-enterprise/alfresco/HEAD/root@18858 c4b6b30b-aa2e-2d43-bbcb-ca4b014f7261
2010-02-25 20:07:09 +00:00

179 lines
5.9 KiB
Java

/*
* Copyright (C) 2009-2010 Alfresco Software Limited.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version 2
* of the License, or (at your option) any later version.
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
* As a special exception to the terms and conditions of version 2.0 of
* the GPL, you may redistribute this Program in connection with Free/Libre
* and Open Source Software ("FLOSS") applications as described in Alfresco's
* FLOSS exception. You should have received a copy of the text describing
* the FLOSS exception, and it is also available here:
* http://www.alfresco.com/legal/licensing"
*/
package org.alfresco.repo.transfer;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.HashSet;
import java.util.LinkedList;
import java.util.List;
import java.util.Queue;
import java.util.Set;
import org.alfresco.service.ServiceRegistry;
import org.alfresco.service.cmr.repository.NodeRef;
import org.alfresco.service.cmr.transfer.NodeFilter;
import org.alfresco.service.cmr.transfer.NodeFinder;
import org.alfresco.service.cmr.transfer.TransferService;
/**
* This class can be used to build a set of node references from a given starting point. The caller can provide a list
* of {@link NodeFinder} objects and a list of {@link NodeFilter} objects. Starting with the nodes supplied by the
* caller, the crawler uses the NodeFinder objects to find other nodes. Each node that is found is then passed to the
* NodeFilter objects to determine whether it should be included or ignored. Any included nodes are then fed back into
* the NodeFinder objects to continue the crawl. This class was originally written to assist users of the
* {@link TransferService} in combination with the {@link ChildAssociatedNodeFinder} and the {@link ContentClassFilter}.
*
* @author brian
*
*/
public class StandardNodeCrawlerImpl
{
private ServiceRegistry serviceRegistry;
private List<NodeFinder> nodeFinders = new ArrayList<NodeFinder>();
private List<NodeFilter> nodeFilters = new ArrayList<NodeFilter>();
/**
*
*/
public StandardNodeCrawlerImpl()
{
super();
}
/**
* @param serviceRegistry
*/
public StandardNodeCrawlerImpl(ServiceRegistry serviceRegistry)
{
super();
this.serviceRegistry = serviceRegistry;
}
/**
* @param nodeService
* the nodeService to set
*/
public void setServiceRegistry(ServiceRegistry serviceRegistry)
{
this.serviceRegistry = serviceRegistry;
}
public Set<NodeRef> crawl(NodeRef... nodes)
{
return crawl(new HashSet<NodeRef>(Arrays.asList(nodes)));
}
public synchronized Set<NodeRef> crawl(Set<NodeRef> startingNodes)
{
init();
Queue<NodeRef> nodesToProcess = new LinkedList<NodeRef>();
nodesToProcess.addAll(startingNodes);
Set<NodeRef> resultingNodeSet = new HashSet<NodeRef>(89);
Set<NodeRef> processedNodes = new HashSet<NodeRef>(89);
// Do we have any more nodes to process?
while (nodesToProcess.peek() != null)
{
// Yes, we do. Read the next noderef from the queue.
NodeRef thisNode = nodesToProcess.poll();
// Check that we haven't already processed it. Skip it if we have, process it if we haven't
if (!processedNodes.contains(thisNode))
{
// Record the fact that we're processing this node
processedNodes.add(thisNode);
// We check this node against any filters that are in place (the nodes
// that we were given to start with are always processed)
if (startingNodes.contains(thisNode) || includeNode(thisNode))
{
resultingNodeSet.add(thisNode);
Set<NodeRef> subsequentNodes = findSubsequentNodes(thisNode);
for (NodeRef node : subsequentNodes)
{
nodesToProcess.add(node);
}
}
}
}
return resultingNodeSet;
}
/**
*
*/
private void init()
{
for (NodeFinder nodeFinder : this.nodeFinders)
{
nodeFinder.setServiceRegistry(serviceRegistry);
nodeFinder.init();
}
for (NodeFilter nodeFilter : this.nodeFilters)
{
nodeFilter.setServiceRegistry(serviceRegistry);
nodeFilter.init();
}
}
/**
* @param thisNode
* @return
*/
private Set<NodeRef> findSubsequentNodes(NodeRef thisNode)
{
Set<NodeRef> foundNodes = new HashSet<NodeRef>(89);
for (NodeFinder finder : nodeFinders)
{
foundNodes.addAll(finder.findFrom(thisNode));
}
return foundNodes;
}
/**
* @param thisNode
* @return
*/
private boolean includeNode(NodeRef thisNode)
{
boolean include = true;
for (int i = 0; include && (i < nodeFilters.size()); ++i)
{
include &= nodeFilters.get(i).accept(thisNode);
}
return include;
}
public synchronized void setNodeFinders(NodeFinder... finders)
{
nodeFinders = Arrays.asList(finders);
}
public synchronized void setNodeFilters(NodeFilter... filters)
{
nodeFilters = Arrays.asList(filters);
}
}