Merged V2.2 to HEAD

8372: Merged V2.1 to V2.2
      8314: Merged V2.0 to V2.1
         7750: Fix for ACT-475: ContentStoreCleaner causes OutOfMemoryError
      8332: Made content URL column larger to accommodate the extra locale info present in 2.1
      8334: Build fix: V2.1 tighter on authentication for getTempWriter
   8376: Merged V2.1 to V2.2
      8325: Fix for AWC-1089
      8361: Workaround for WCM-882: All metadata extracters can now handle zero length files


git-svn-id: https://svn.alfresco.com/repos/alfresco-enterprise/alfresco/HEAD/root@8497 c4b6b30b-aa2e-2d43-bbcb-ca4b014f7261
This commit is contained in:
Derek Hulley
2008-03-11 06:22:28 +00:00
parent ceed05d26f
commit cda4e6105f
33 changed files with 1102 additions and 246 deletions

View File

@@ -63,6 +63,7 @@
<ref bean="patch.db-V2.2-4-ExtraIndexesAndConstraints" /> <ref bean="patch.db-V2.2-4-ExtraIndexesAndConstraints" />
<ref bean="patch.db-V2.2-5-MappedFKIndexes" /> <ref bean="patch.db-V2.2-5-MappedFKIndexes" />
<ref bean="patch.db-V1.4-TxnCommitTimeIndex" /> <ref bean="patch.db-V1.4-TxnCommitTimeIndex" />
<ref bean="patch.db-V2.0-ContentUrls" />
<ref bean="patch.db-V2.1-JBPMData" /> <ref bean="patch.db-V2.1-JBPMData" />
<ref bean="patch.db-V2.1-VersionColumns2" /> <ref bean="patch.db-V2.1-VersionColumns2" />
<ref bean="patch.db-V2.1-JBPMProcessKey" /> <ref bean="patch.db-V2.1-JBPMProcessKey" />

View File

@@ -34,6 +34,9 @@
<property name="avmNodeDAO"> <property name="avmNodeDAO">
<ref bean="avmNodeDAO"/> <ref bean="avmNodeDAO"/>
</property> </property>
<property name="contentUrlDAO">
<ref bean="contentUrlDAO"/>
</property>
<property name="transactionService" > <property name="transactionService" >
<ref bean="transactionService" /> <ref bean="transactionService" />
</property> </property>

View File

@@ -0,0 +1,24 @@
--
-- Title: Indexes for alf_content_url table
-- Database: Generic
-- Since: V2.0 Schema 44
-- Author: Derek Hulley
--
-- Please contact support@alfresco.com if you need assistance with the upgrade.
--
-- Content URLs
SELECT COUNT(*) FROM alf_content_url;
CREATE INDEX idx_alf_con_urls ON alf_content_url (content_url);(optional)
--
-- Record script finish
--
DELETE FROM alf_applied_patch WHERE id = 'patch.db-V2.0-ContentUrls';
INSERT INTO alf_applied_patch
(id, description, fixes_from_schema, fixes_to_schema, applied_to_schema, target_schema, applied_on_date, applied_to_server, was_executed, succeeded, report)
VALUES
(
'patch.db-V2.0-ContentUrls', 'Manually executed script upgrade V2.0: Indexes for alf_content_url table',
0, 43, -1, 44, null, 'UNKOWN', 1, 1, 'Script completed'
);

View File

@@ -0,0 +1,24 @@
--
-- Title: Indexes for alf_content_url table
-- Database: PostgreSQL
-- Since: V2.0 Schema 44
-- Author: Derek Hulley
--
-- Please contact support@alfresco.com if you need assistance with the upgrade.
--
-- Content URLs
SELECT COUNT(*) FROM alf_content_url;
CREATE INDEX idx_alf_con_urls ON alf_content_url (content_url);(optional)
--
-- Record script finish
--
DELETE FROM alf_applied_patch WHERE id = 'patch.db-V2.0-ContentUrls';
INSERT INTO alf_applied_patch
(id, description, fixes_from_schema, fixes_to_schema, applied_to_schema, target_schema, applied_on_date, applied_to_server, was_executed, succeeded, report)
VALUES
(
'patch.db-V2.0-ContentUrls', 'Manually executed script upgrade V2.0: Indexes for alf_content_url table',
0, 43, -1, 44, null, 'UNKOWN', TRUE, TRUE, 'Script completed'
);

View File

@@ -54,6 +54,9 @@
<!-- TODO: Move into org/alfresco/repo/domain/hibernate/ --> <!-- TODO: Move into org/alfresco/repo/domain/hibernate/ -->
<value>org/alfresco/repo/audit/hibernate/Audit.hbm.xml</value> <value>org/alfresco/repo/audit/hibernate/Audit.hbm.xml</value>
<!-- Content URL config -->
<value>org/alfresco/repo/domain/hibernate/ContentUrl.hbm.xml</value>
<!-- --> <!-- -->
<!-- JBoss jBPM Workflow Engine --> <!-- JBoss jBPM Workflow Engine -->
<!-- --> <!-- -->
@@ -174,6 +177,8 @@
<prop key="org.alfresco.repo.audit.hibernate.AuditConfigImpl">${cache.strategy}</prop> <prop key="org.alfresco.repo.audit.hibernate.AuditConfigImpl">${cache.strategy}</prop>
<prop key="org.alfresco.repo.audit.hibernate.AuditDateImpl">${cache.strategy}</prop> <prop key="org.alfresco.repo.audit.hibernate.AuditDateImpl">${cache.strategy}</prop>
<prop key="org.alfresco.repo.audit.hibernate.AuditSourceImpl">${cache.strategy}</prop> <prop key="org.alfresco.repo.audit.hibernate.AuditSourceImpl">${cache.strategy}</prop>
<prop key="org.alfresco.repo.domain.hibernate.ContentUrlImpl">${cache.strategy}</prop>
</props> </props>
</property> </property>
<property name="collectionCacheStrategies" > <property name="collectionCacheStrategies" >
@@ -342,4 +347,23 @@
</property> </property>
</bean> </bean>
<bean id="contentUrlDAO" class="org.springframework.aop.framework.ProxyFactoryBean">
<property name="proxyInterfaces">
<value>org.alfresco.repo.domain.ContentUrlDAO</value>
</property>
<property name="target">
<ref bean="contentUrlDAOImpl" />
</property>
<property name="interceptorNames">
<list>
<value>sessionSizeResourceInterceptor</value>
</list>
</property>
</bean>
<bean id="contentUrlDAOImpl" class="org.alfresco.repo.domain.hibernate.HibernateContentUrlDAOImpl">
<property name="sessionFactory">
<ref bean="sessionFactory"/>
</property>
</bean>
</beans> </beans>

View File

@@ -1407,4 +1407,20 @@
</property> </property>
</bean> </bean>
<bean id="patch.db-V2.0-ContentUrls" class="org.alfresco.repo.admin.patch.impl.SchemaUpgradeScriptPatch" parent="basePatch">
<property name="id"><value>patch.db-V2.0-ContentUrls</value></property>
<property name="description"><value>patch.schemaUpgradeScript.description</value></property>
<property name="fixesFromSchema"><value>0</value></property>
<property name="fixesToSchema"><value>123</value></property>
<property name="targetSchema"><value>124</value></property>
<property name="scriptUrl">
<value>classpath:alfresco/dbscripts/upgrade/2.0/${db.script.dialect}/AlfrescoSchemaUpdate-2.0-ContentUrls.sql</value>
</property>
<property name="dependsOn" >
<list>
<ref bean="patch.db-V2.2-4-ExtraIndexesAndConstraints" />
</list>
</property>
</bean>
</beans> </beans>

View File

@@ -19,4 +19,4 @@ version.build=@build-number@
# Schema number # Schema number
version.schema=123 version.schema=124

View File

@@ -83,10 +83,21 @@ public interface AVMNodeDAO
public List<AVMNode> getOrphans(int batchSize); public List<AVMNode> getOrphans(int batchSize);
/** /**
* Get all content urls in he AVM Repository. * Get all content urls in the AVM Repository.
* @return A List of URL Strings. * @param contentUrlHandler the handler that will be called with the URLs
*/ */
public List<String> getContentUrls(); public void getContentUrls(ContentUrlHandler handler);
/**
* A callback handler for iterating over the content URLs
*
* @author Derek Hulley
* @since 2.0
*/
public interface ContentUrlHandler
{
void handle(String contentUrl);
}
/** /**
* Get all the nodes that are new in the given store. * Get all the nodes that are new in the given store.

View File

@@ -38,6 +38,8 @@ import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
import org.hibernate.CacheMode; import org.hibernate.CacheMode;
import org.hibernate.Query; import org.hibernate.Query;
import org.hibernate.ScrollMode;
import org.hibernate.ScrollableResults;
import org.springframework.orm.hibernate3.support.HibernateDaoSupport; import org.springframework.orm.hibernate3.support.HibernateDaoSupport;
/** /**
@@ -157,10 +159,15 @@ class AVMNodeDAOHibernate extends HibernateDaoSupport implements
* @return A List of URL Strings. * @return A List of URL Strings.
*/ */
@SuppressWarnings("unchecked") @SuppressWarnings("unchecked")
public List<String> getContentUrls() public void getContentUrls(ContentUrlHandler handler)
{ {
Query query = getSession().getNamedQuery("PlainFileNode.GetContentUrls"); Query query = getSession().getNamedQuery("PlainFileNode.GetContentUrls");
return (List<String>)query.list(); ScrollableResults results = query.scroll(ScrollMode.FORWARD_ONLY);
while (results.next())
{
String contentUrl = results.getText(0);
handler.handle(contentUrl);
}
} }
/** /**

View File

@@ -25,7 +25,6 @@
package org.alfresco.repo.content; package org.alfresco.repo.content;
import java.util.Date; import java.util.Date;
import java.util.Set;
import org.alfresco.service.cmr.repository.ContentIOException; import org.alfresco.service.cmr.repository.ContentIOException;
import org.alfresco.service.cmr.repository.ContentReader; import org.alfresco.service.cmr.repository.ContentReader;
@@ -150,23 +149,20 @@ public abstract class AbstractContentStore implements ContentStore
} }
/** /**
* Searches for URLs using null dates. * @see #getUrls(Date, Date, ContentUrlHandler)
*
* @see ContentStore#getUrls(java.util.Date, java.util.Date)
*/ */
public final Set<String> getUrls() public final void getUrls(ContentUrlHandler handler) throws ContentIOException
{ {
return getUrls(null, null); getUrls(null, null, handler);
} }
/** /**
* Override if the derived class supports the operation. * Override to provide an implementation. If no implementation is supplied, then the store will not support
* cleaning of orphaned content binaries.
* *
* @throws UnsupportedOperationException always * @throws UnsupportedOperationException always
*
* @since 2.1
*/ */
public Set<String> getUrls(Date createdAfter, Date createdBefore) public void getUrls(Date createdAfter, Date createdBefore, ContentUrlHandler handler) throws ContentIOException
{ {
throw new UnsupportedOperationException(); throw new UnsupportedOperationException();
} }
@@ -256,6 +252,16 @@ public abstract class AbstractContentStore implements ContentStore
return writer; return writer;
} }
/**
* @see ContentContext
* @see ContentStore#getWriter(ContentContext)
*/
public final ContentWriter getWriter(ContentReader existingContentReader, String newContentUrl)
{
ContentContext ctx = new ContentContext(existingContentReader, newContentUrl);
return getWriter(ctx);
}
/** /**
* Simple implementation that uses the * Simple implementation that uses the
* {@link ContentReader#exists() reader's exists} method as its implementation. * {@link ContentReader#exists() reader's exists} method as its implementation.
@@ -266,14 +272,4 @@ public abstract class AbstractContentStore implements ContentStore
ContentReader reader = getReader(contentUrl); ContentReader reader = getReader(contentUrl);
return reader.exists(); return reader.exists();
} }
/**
* @see ContentContext
* @see ContentStore#getWriter(ContentContext)
*/
public final ContentWriter getWriter(ContentReader existingContentReader, String newContentUrl)
{
ContentContext ctx = new ContentContext(existingContentReader, newContentUrl);
return getWriter(ctx);
}
} }

View File

@@ -27,12 +27,14 @@ package org.alfresco.repo.content;
import java.nio.ByteBuffer; import java.nio.ByteBuffer;
import java.nio.channels.FileChannel; import java.nio.channels.FileChannel;
import java.nio.channels.ReadableByteChannel; import java.nio.channels.ReadableByteChannel;
import java.util.HashSet;
import java.util.Set; import java.util.Set;
import javax.transaction.UserTransaction; import javax.transaction.UserTransaction;
import junit.framework.TestCase; import junit.framework.TestCase;
import org.alfresco.repo.content.ContentStore.ContentUrlHandler;
import org.alfresco.service.cmr.repository.ContentReader; import org.alfresco.service.cmr.repository.ContentReader;
import org.alfresco.service.transaction.TransactionService; import org.alfresco.service.transaction.TransactionService;
import org.alfresco.util.ApplicationContextHelper; import org.alfresco.util.ApplicationContextHelper;
@@ -118,7 +120,18 @@ public abstract class AbstractReadOnlyContentStoreTest extends TestCase
ContentStore store = getStore(); ContentStore store = getStore();
try try
{ {
Set<String> contentUrls = store.getUrls(); final Set<String> contentUrls = new HashSet<String>(5);
ContentUrlHandler handler = new ContentUrlHandler()
{
public void handle(String contentUrl)
{
if (contentUrls.size() < 50)
{
contentUrls.add(contentUrl);
}
}
};
store.getUrls(handler);
if (contentUrls.size() > 0) if (contentUrls.size() > 0)
{ {
return (String) contentUrls.toArray()[0]; return (String) contentUrls.toArray()[0];

View File

@@ -25,9 +25,7 @@
package org.alfresco.repo.content; package org.alfresco.repo.content;
import java.util.Date; import java.util.Date;
import java.util.HashSet;
import java.util.List; import java.util.List;
import java.util.Set;
import java.util.concurrent.locks.ReentrantReadWriteLock; import java.util.concurrent.locks.ReentrantReadWriteLock;
import java.util.concurrent.locks.ReentrantReadWriteLock.ReadLock; import java.util.concurrent.locks.ReentrantReadWriteLock.ReadLock;
import java.util.concurrent.locks.ReentrantReadWriteLock.WriteLock; import java.util.concurrent.locks.ReentrantReadWriteLock.WriteLock;
@@ -357,50 +355,38 @@ public abstract class AbstractRoutingContentStore implements ContentStore
return writer; return writer;
} }
/**
* @see
*/
public ContentWriter getWriter(ContentReader existingContentReader, String newContentUrl) throws ContentIOException public ContentWriter getWriter(ContentReader existingContentReader, String newContentUrl) throws ContentIOException
{ {
return getWriter(new ContentContext(existingContentReader, newContentUrl)); return getWriter(new ContentContext(existingContentReader, newContentUrl));
} }
/** /**
* Compile a set of URLs from all stores. * @see #getUrls(Date, Date, ContentUrlHandler)
*/ */
public Set<String> getUrls() throws ContentIOException public void getUrls(ContentUrlHandler handler) throws ContentIOException
{ {
Set<String> urls = new HashSet<String>(1139); getUrls(null, null, handler);
List<ContentStore> stores = getAllStores();
for (ContentStore store : stores)
{
Set<String> storeUrls = store.getUrls();
urls.addAll(storeUrls);
}
if (logger.isDebugEnabled())
{
logger.debug("Found " + urls.size() + " URLs from " + stores.size() + " stores");
}
return urls;
} }
/** /**
* Compile a set of URLs from all stores given the date range. * Passes the call to each of the stores wrapped by this store
*
* @see ContentStore#getUrls(Date, Date, ContentUrlHandler)
*/ */
public Set<String> getUrls(Date createdAfter, Date createdBefore) throws ContentIOException public void getUrls(Date createdAfter, Date createdBefore, ContentUrlHandler handler) throws ContentIOException
{ {
Set<String> urls = new HashSet<String>(1139);
List<ContentStore> stores = getAllStores(); List<ContentStore> stores = getAllStores();
for (ContentStore store : stores) for (ContentStore store : stores)
{ {
Set<String> storeUrls = store.getUrls(createdAfter, createdBefore); try
urls.addAll(storeUrls);
}
if (logger.isDebugEnabled())
{ {
logger.debug("Found " + urls.size() + " URLs from " + stores.size() + " stores"); store.getUrls(createdAfter, createdBefore, handler);
}
catch (UnsupportedOperationException e)
{
// Support of this is not mandatory
}
} }
return urls;
} }
/** /**

View File

@@ -36,8 +36,8 @@ import java.nio.channels.FileChannel;
import java.nio.channels.ReadableByteChannel; import java.nio.channels.ReadableByteChannel;
import java.util.Date; import java.util.Date;
import java.util.Locale; import java.util.Locale;
import java.util.Set;
import org.alfresco.repo.content.ContentStore.ContentUrlHandler;
import org.alfresco.service.cmr.repository.ContentIOException; import org.alfresco.service.cmr.repository.ContentIOException;
import org.alfresco.service.cmr.repository.ContentReader; import org.alfresco.service.cmr.repository.ContentReader;
import org.alfresco.service.cmr.repository.ContentStreamListener; import org.alfresco.service.cmr.repository.ContentStreamListener;
@@ -329,23 +329,34 @@ public abstract class AbstractWritableContentStoreTest extends AbstractReadOnlyC
assertTrue("After-content reader should be closed after reading", readerAfterWrite.isClosed()); assertTrue("After-content reader should be closed after reading", readerAfterWrite.isClosed());
} }
/**
* Helper method to check if a store contains a particular URL using the getUrl method
*/
private boolean searchForUrl(ContentStore store, final String contentUrl, Date from, Date to)
{
final boolean[] found = new boolean[] {false};
ContentUrlHandler handler = new ContentUrlHandler()
{
public void handle(String checkContentUrl)
{
if (contentUrl.equals(checkContentUrl))
{
found[0] = true;
}
}
};
getStore().getUrls(from, to, handler);
return found[0];
}
public void testGetUrls() public void testGetUrls()
{ {
ContentStore store = getStore();
try
{
store.getUrls();
}
catch (UnsupportedOperationException e)
{
logger.warn("Store test " + getName() + " not possible on " + store.getClass().getName());
return;
}
ContentWriter writer = getWriter(); ContentWriter writer = getWriter();
writer.putContent("Content for " + getName()); writer.putContent("Content for " + getName());
Set<String> contentUrls = store.getUrls(); final String contentUrl = writer.getContentUrl();
String contentUrl = writer.getContentUrl(); ContentStore store = getStore();
assertTrue("New content not found in URL set", contentUrls.contains(contentUrl)); boolean inStore = searchForUrl(store, contentUrl, null, null);
assertTrue("New content not found in URL set", inStore);
} }
public void testDeleteSimple() throws Exception public void testDeleteSimple() throws Exception
@@ -603,7 +614,7 @@ public abstract class AbstractWritableContentStoreTest extends AbstractReadOnlyC
// Ensure that this test can be done // Ensure that this test can be done
try try
{ {
store.getUrls(); searchForUrl(store, "abc", null, null);
} }
catch (UnsupportedOperationException e) catch (UnsupportedOperationException e)
{ {
@@ -612,10 +623,10 @@ public abstract class AbstractWritableContentStoreTest extends AbstractReadOnlyC
} }
// Proceed with the test // Proceed with the test
ContentWriter writer = getWriter(); ContentWriter writer = getWriter();
Set<String> contentUrls = store.getUrls();
String contentUrl = writer.getContentUrl(); String contentUrl = writer.getContentUrl();
assertTrue("Writer URL not listed by store", contentUrls.contains(contentUrl));
boolean inStore = searchForUrl(store, contentUrl, null, null);
assertTrue("Writer URL not listed by store", inStore);
Date yesterday = new Date(System.currentTimeMillis() - 3600L * 1000L * 24L); Date yesterday = new Date(System.currentTimeMillis() - 3600L * 1000L * 24L);
@@ -623,12 +634,12 @@ public abstract class AbstractWritableContentStoreTest extends AbstractReadOnlyC
writer.putContent("The quick brown fox..."); writer.putContent("The quick brown fox...");
// check again // check again
contentUrls = store.getUrls(); inStore = searchForUrl(store, contentUrl, null, null);
assertTrue("Writer URL not listed by store", contentUrls.contains(contentUrl)); assertTrue("Writer URL not listed by store", inStore);
// check that the query for content created before this time yesterday doesn't return the URL // check that the query for content created before this time yesterday doesn't return the URL
contentUrls = store.getUrls(null, yesterday); inStore = searchForUrl(store, contentUrl, null, yesterday);
assertFalse("URL was younger than required, but still shows up", contentUrls.contains(contentUrl)); assertFalse("URL was younger than required, but still shows up", inStore);
} }
/** /**

View File

@@ -25,7 +25,6 @@
package org.alfresco.repo.content; package org.alfresco.repo.content;
import java.util.Date; import java.util.Date;
import java.util.Set;
import org.alfresco.service.cmr.repository.ContentAccessor; import org.alfresco.service.cmr.repository.ContentAccessor;
import org.alfresco.service.cmr.repository.ContentIOException; import org.alfresco.service.cmr.repository.ContentIOException;
@@ -195,24 +194,23 @@ public interface ContentStore
* *
* @see #getUrls(Date, Date) * @see #getUrls(Date, Date)
*/ */
public Set<String> getUrls(); public void getUrls(ContentUrlHandler handler) throws ContentIOException;
/** /**
* Get a set of all content URLs in the store. This indicates all content * Get a set of all content URLs in the store. This indicates all content available for reads.
* available for reads.
* *
* @param createdAfter * @param createdAfter
* all URLs returned must have been created after this date. May be null. * all URLs returned must have been created after this date. May be null.
* @param createdBefore * @param createdBefore
* all URLs returned must have been created before this date. May be null. * all URLs returned must have been created before this date. May be null.
* @return * @param handler
* Returns a complete set of the unique URLs of all available content in the store * the callback that will passed each URL
* @throws ContentIOException
* if an error occurs
* @throws UnsupportedOperationException * @throws UnsupportedOperationException
* if the store is unable to provide the information * if the store is unable to provide the information
* @throws ContentIOException
* if an IO error occurs
*/ */
public Set<String> getUrls(Date createdAfter, Date createdBefore); public void getUrls(Date createdAfter, Date createdBefore, ContentUrlHandler handler) throws ContentIOException;
/** /**
* Deletes the content at the given URL. * Deletes the content at the given URL.
@@ -229,8 +227,19 @@ public interface ContentStore
* if the store is unable to perform the action * if the store is unable to perform the action
* @throws UnsupportedContentUrlException * @throws UnsupportedContentUrlException
* if the content URL supplied is not supported by the store * if the content URL supplied is not supported by the store
* @throws ContentIOException * @throws ContentIOException if an error occurs
* if an IO error occurs * if an IO error occurs
*/ */
public boolean delete(String contentUrl); public boolean delete(String contentUrl);
/**
* Iterface for to use during iteration over content URLs.
*
* @author Derek Hulley
* @since 2.0
*/
public interface ContentUrlHandler
{
void handle(String contentUrl);
}
} }

View File

@@ -27,6 +27,7 @@ package org.alfresco.repo.content;
import java.io.File; import java.io.File;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Collections; import java.util.Collections;
import java.util.Date;
import java.util.List; import java.util.List;
import net.sf.ehcache.Cache; import net.sf.ehcache.Cache;
@@ -34,6 +35,7 @@ import net.sf.ehcache.CacheManager;
import org.alfresco.repo.cache.EhCacheAdapter; import org.alfresco.repo.cache.EhCacheAdapter;
import org.alfresco.repo.content.filestore.FileContentStore; import org.alfresco.repo.content.filestore.FileContentStore;
import org.alfresco.service.cmr.repository.ContentIOException;
import org.alfresco.service.cmr.repository.ContentReader; import org.alfresco.service.cmr.repository.ContentReader;
import org.alfresco.service.cmr.repository.ContentWriter; import org.alfresco.service.cmr.repository.ContentWriter;
import org.alfresco.util.TempFileProvider; import org.alfresco.util.TempFileProvider;
@@ -217,6 +219,12 @@ public class RoutingContentStoreTest extends AbstractWritableContentStoreTest
{ {
return fileStore.getReader(contentUrl); return fileStore.getReader(contentUrl);
} }
@Override
public void getUrls(Date createdAfter, Date createdBefore, ContentUrlHandler handler) throws ContentIOException
{
fileStore.getUrls(createdAfter, createdBefore, handler);
}
} }
/** /**
@@ -239,5 +247,11 @@ public class RoutingContentStoreTest extends AbstractWritableContentStoreTest
{ {
throw new UnsupportedContentUrlException(this, contentUrl); throw new UnsupportedContentUrlException(this, contentUrl);
} }
@Override
public void getUrls(Date createdAfter, Date createdBefore, ContentUrlHandler handler) throws ContentIOException
{
throw new UnsupportedOperationException("getUrls not supported");
}
} }
} }

View File

@@ -20,31 +20,31 @@
* and Open Source Software ("FLOSS") applications as described in Alfresco's * and Open Source Software ("FLOSS") applications as described in Alfresco's
* FLOSS exception. You should have recieved a copy of the text describing * FLOSS exception. You should have recieved a copy of the text describing
* the FLOSS exception, and it is also available here: * the FLOSS exception, and it is also available here:
* http://www.alfresco.com/legal/licensing" * http://www.alfresco.com/legal/licensing
*/ */
package org.alfresco.repo.content.cleanup; package org.alfresco.repo.content.cleanup;
import java.io.Serializable; import java.io.Serializable;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Date; import java.util.Date;
import java.util.HashSet;
import java.util.List; import java.util.List;
import java.util.Set;
import org.alfresco.error.AlfrescoRuntimeException; import org.alfresco.error.AlfrescoRuntimeException;
import org.alfresco.repo.avm.AVMNodeDAO; import org.alfresco.repo.avm.AVMNodeDAO;
import org.alfresco.repo.avm.AVMNodeDAO.ContentUrlHandler;
import org.alfresco.repo.content.ContentStore; import org.alfresco.repo.content.ContentStore;
import org.alfresco.repo.domain.ContentUrlDAO;
import org.alfresco.repo.domain.Node;
import org.alfresco.repo.node.db.NodeDaoService; import org.alfresco.repo.node.db.NodeDaoService;
import org.alfresco.repo.node.db.NodeDaoService.NodePropertyHandler;
import org.alfresco.repo.transaction.RetryingTransactionHelper; import org.alfresco.repo.transaction.RetryingTransactionHelper;
import org.alfresco.repo.transaction.RetryingTransactionHelper.RetryingTransactionCallback; import org.alfresco.repo.transaction.RetryingTransactionHelper.RetryingTransactionCallback;
import org.alfresco.service.cmr.dictionary.DataTypeDefinition; import org.alfresco.service.cmr.dictionary.DataTypeDefinition;
import org.alfresco.service.cmr.dictionary.DictionaryService; import org.alfresco.service.cmr.dictionary.DictionaryService;
import org.alfresco.service.cmr.repository.ContentData; import org.alfresco.service.cmr.repository.ContentData;
import org.alfresco.service.cmr.repository.ContentIOException; import org.alfresco.service.cmr.repository.datatype.DefaultTypeConverter;
import org.alfresco.service.cmr.repository.ContentReader;
import org.alfresco.service.transaction.TransactionService; import org.alfresco.service.transaction.TransactionService;
import org.alfresco.util.PropertyCheck; import org.alfresco.util.PropertyCheck;
import org.alfresco.util.VmShutdownListener;
import org.apache.commons.logging.Log; import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
@@ -60,12 +60,11 @@ public class ContentStoreCleaner
{ {
private static Log logger = LogFactory.getLog(ContentStoreCleaner.class); private static Log logger = LogFactory.getLog(ContentStoreCleaner.class);
private static VmShutdownListener vmShutdownListener = new VmShutdownListener(ContentStoreCleaner.class.getName());
private DictionaryService dictionaryService; private DictionaryService dictionaryService;
private NodeDaoService nodeDaoService; private NodeDaoService nodeDaoService;
private TransactionService transactionService;
private AVMNodeDAO avmNodeDAO; private AVMNodeDAO avmNodeDAO;
private ContentUrlDAO contentUrlDAO;
private TransactionService transactionService;
private List<ContentStore> stores; private List<ContentStore> stores;
private List<ContentStoreCleanerListener> listeners; private List<ContentStoreCleanerListener> listeners;
private int protectDays; private int protectDays;
@@ -94,7 +93,6 @@ public class ContentStoreCleaner
} }
/** /**
* Setter for Spring.
* @param avmNodeDAO The AVM Node DAO to get urls with. * @param avmNodeDAO The AVM Node DAO to get urls with.
*/ */
public void setAvmNodeDAO(AVMNodeDAO avmNodeDAO) public void setAvmNodeDAO(AVMNodeDAO avmNodeDAO)
@@ -102,6 +100,14 @@ public class ContentStoreCleaner
this.avmNodeDAO = avmNodeDAO; this.avmNodeDAO = avmNodeDAO;
} }
/**
* @param contentUrlDAO DAO for recording valid <b>Content URLs</b>
*/
public void setContentUrlDAO(ContentUrlDAO contentUrlDAO)
{
this.contentUrlDAO = contentUrlDAO;
}
/** /**
* @param transactionService the component to ensure proper transactional wrapping * @param transactionService the component to ensure proper transactional wrapping
*/ */
@@ -144,6 +150,8 @@ public class ContentStoreCleaner
{ {
PropertyCheck.mandatory(this, "dictionaryService", dictionaryService); PropertyCheck.mandatory(this, "dictionaryService", dictionaryService);
PropertyCheck.mandatory(this, "nodeDaoService", nodeDaoService); PropertyCheck.mandatory(this, "nodeDaoService", nodeDaoService);
PropertyCheck.mandatory(this, "avmNodeDAO", avmNodeDAO);
PropertyCheck.mandatory(this, "contentUrlDAO", contentUrlDAO);
PropertyCheck.mandatory(this, "transactionService", transactionService); PropertyCheck.mandatory(this, "transactionService", transactionService);
PropertyCheck.mandatory(this, "listeners", listeners); PropertyCheck.mandatory(this, "listeners", listeners);
@@ -160,126 +168,124 @@ public class ContentStoreCleaner
} }
} }
private Set<String> getValidUrls() private void removeContentUrlsPresentInMetadata()
{ {
RetryingTransactionHelper txnHelper = transactionService.getRetryingTransactionHelper(); RetryingTransactionHelper txnHelper = transactionService.getRetryingTransactionHelper();
final DataTypeDefinition contentDataType = dictionaryService.getDataType(DataTypeDefinition.CONTENT); // Remove all the Content URLs for the ADM repository
// wrap to make the request in a transaction // Handler that records the URL
RetryingTransactionCallback<List<Serializable>> getUrlsCallback = new RetryingTransactionCallback<List<Serializable>>() final NodePropertyHandler nodePropertyHandler = new NodePropertyHandler()
{ {
public List<Serializable> execute() throws Throwable public void handle(Node node, Serializable value)
{ {
return nodeDaoService.getPropertyValuesByActualType(contentDataType); // Convert the values to ContentData and extract the URLs
ContentData contentData = DefaultTypeConverter.INSTANCE.convert(ContentData.class, value);
String contentUrl = contentData.getContentUrl();
if (contentUrl != null)
{
contentUrlDAO.deleteContentUrl(contentUrl);
}
} }
}; };
// execute in READ-ONLY txn final DataTypeDefinition contentDataType = dictionaryService.getDataType(DataTypeDefinition.CONTENT);
List<Serializable> values = txnHelper.doInTransaction(getUrlsCallback, true); // execute in READ-WRITE txn
RetryingTransactionCallback<Object> getUrlsCallback = new RetryingTransactionCallback<Object>()
{
public Object execute() throws Exception
{
nodeDaoService.getPropertyValuesByActualType(contentDataType, nodePropertyHandler);
return null;
};
};
txnHelper.doInTransaction(getUrlsCallback);
// Do the same for the AVM repository. // Do the same for the AVM repository.
RetryingTransactionCallback<List<String>> getAVMUrlsCallback = new RetryingTransactionCallback<List<String>>() final ContentUrlHandler handler = new ContentUrlHandler()
{ {
public List<String> execute() throws Exception public void handle(String contentUrl)
{ {
return avmNodeDAO.getContentUrls(); contentUrlDAO.deleteContentUrl(contentUrl);
} }
}; };
// execute in READ-ONLY txn // execute in READ-WRITE txn
List<String> avmContentUrls = txnHelper.doInTransaction(getAVMUrlsCallback, true); RetryingTransactionCallback<Object> getAVMUrlsCallback = new RetryingTransactionCallback<Object>()
// get all valid URLs
Set<String> validUrls = new HashSet<String>(values.size());
// convert the strings to objects and extract the URL
for (Serializable value : values)
{ {
ContentData contentData = (ContentData) value; public Object execute() throws Exception
if (contentData.getContentUrl() != null)
{ {
// a URL was present avmNodeDAO.getContentUrls(handler);
validUrls.add(contentData.getContentUrl()); return null;
} }
} };
// put all the avm urls into validUrls. txnHelper.doInTransaction(getAVMUrlsCallback);
for (String url : avmContentUrls)
{
validUrls.add(url);
} }
// done private void addContentUrlsPresentInStores()
if (logger.isDebugEnabled())
{ {
logger.debug("Found " + validUrls.size() + " valid URLs in metadata"); org.alfresco.repo.content.ContentStore.ContentUrlHandler handler = new org.alfresco.repo.content.ContentStore.ContentUrlHandler()
{
public void handle(String contentUrl)
{
contentUrlDAO.createContentUrl(contentUrl);
}
};
Date checkAllBeforeDate = new Date(System.currentTimeMillis() - (long) protectDays * 3600L * 1000L * 24L);
for (ContentStore store : stores)
{
store.getUrls(null, checkAllBeforeDate, handler);
} }
return validUrls;
} }
public void execute() public void execute()
{ {
checkProperties(); checkProperties();
try
{
Set<String> validUrls = getValidUrls();
// now clean each store in turn
for (ContentStore store : stores)
{
try
{
clean(validUrls, store);
}
catch (UnsupportedOperationException e)
{
throw new ContentIOException(
"Unable to clean store as the necessary operations are not supported: " + store,
e);
}
}
}
catch (ContentIOException e)
{
throw e;
}
catch (Throwable e)
{
// If the VM is shutting down, then ignore
if (vmShutdownListener.isVmShuttingDown())
{
// Ignore
}
else
{
logger.error("Exception during cleanup of content", e);
}
}
}
private void clean(Set<String> validUrls, ContentStore store)
{
Date checkAllBeforeDate = new Date(System.currentTimeMillis() - (long) protectDays * 3600L * 1000L * 24L);
// get the store's URLs
Set<String> storeUrls = store.getUrls(null, checkAllBeforeDate);
// remove all URLs that occur in the validUrls
storeUrls.removeAll(validUrls);
// now clean the store
for (String url : storeUrls)
{
ContentReader sourceReader = store.getReader(url);
// announce this to the listeners
for (ContentStoreCleanerListener listener : listeners)
{
// get a fresh reader
ContentReader listenerReader = sourceReader.getReader();
// call it
listener.beforeDelete(listenerReader);
}
// delete it
store.delete(url);
if (logger.isDebugEnabled()) if (logger.isDebugEnabled())
{ {
logger.debug("Removed URL from store: \n" + logger.debug("Starting content store cleanup.");
" Store: " + store + "\n" +
" URL: " + url);
} }
// This handler removes the URLs from all the stores
final org.alfresco.repo.domain.ContentUrlDAO.ContentUrlHandler handler = new org.alfresco.repo.domain.ContentUrlDAO.ContentUrlHandler()
{
public void handle(String contentUrl)
{
for (ContentStore store : stores)
{
if (logger.isDebugEnabled())
{
if (store.isWriteSupported())
{
logger.debug(" Deleting content URL: " + contentUrl);
}
}
store.delete(contentUrl);
}
}
};
// execute in READ-WRITE txn
RetryingTransactionCallback<Object> executeCallback = new RetryingTransactionCallback<Object>()
{
public Object execute() throws Exception
{
// Delete all the URLs
contentUrlDAO.deleteAllContentUrls();
// Populate the URLs from the content stores
addContentUrlsPresentInStores();
// Remove URLs present in the metadata
removeContentUrlsPresentInMetadata();
// Any remaining URLs are URls present in the stores but not in the metadata
contentUrlDAO.getAllContentUrls(handler);
// Delete all the URLs
contentUrlDAO.deleteAllContentUrls();
return null;
};
};
transactionService.getRetryingTransactionHelper().doInTransaction(executeCallback);
// Done
if (logger.isDebugEnabled())
{
logger.debug(" Content store cleanup completed.");
} }
} }
} }

View File

@@ -27,10 +27,12 @@ package org.alfresco.repo.content.cleanup;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Collections; import java.util.Collections;
import java.util.List; import java.util.List;
import java.util.concurrent.CountDownLatch;
import org.alfresco.repo.avm.AVMNodeDAO; import org.alfresco.repo.avm.AVMNodeDAO;
import org.alfresco.repo.content.ContentStore; import org.alfresco.repo.content.ContentStore;
import org.alfresco.repo.content.filestore.FileContentStore; import org.alfresco.repo.content.filestore.FileContentStore;
import org.alfresco.repo.domain.ContentUrlDAO;
import org.alfresco.repo.node.db.NodeDaoService; import org.alfresco.repo.node.db.NodeDaoService;
import org.alfresco.service.ServiceRegistry; import org.alfresco.service.ServiceRegistry;
import org.alfresco.service.cmr.dictionary.DictionaryService; import org.alfresco.service.cmr.dictionary.DictionaryService;
@@ -66,6 +68,7 @@ public class ContentStoreCleanerTest extends TestCase
DictionaryService dictionaryService = serviceRegistry.getDictionaryService(); DictionaryService dictionaryService = serviceRegistry.getDictionaryService();
NodeDaoService nodeDaoService = (NodeDaoService) ctx.getBean("nodeDaoService"); NodeDaoService nodeDaoService = (NodeDaoService) ctx.getBean("nodeDaoService");
AVMNodeDAO avmNodeDAO = (AVMNodeDAO) ctx.getBean("avmNodeDAO"); AVMNodeDAO avmNodeDAO = (AVMNodeDAO) ctx.getBean("avmNodeDAO");
ContentUrlDAO contentUrlDAO = (ContentUrlDAO) ctx.getBean("contentUrlDAO");
// we need a store // we need a store
store = new FileContentStore(TempFileProvider.getTempDir().getAbsolutePath()); store = new FileContentStore(TempFileProvider.getTempDir().getAbsolutePath());
@@ -80,6 +83,7 @@ public class ContentStoreCleanerTest extends TestCase
cleaner.setDictionaryService(dictionaryService); cleaner.setDictionaryService(dictionaryService);
cleaner.setNodeDaoService(nodeDaoService); cleaner.setNodeDaoService(nodeDaoService);
cleaner.setAvmNodeDAO(avmNodeDAO); cleaner.setAvmNodeDAO(avmNodeDAO);
cleaner.setContentUrlDAO(contentUrlDAO);
cleaner.setStores(Collections.singletonList(store)); cleaner.setStores(Collections.singletonList(store));
cleaner.setListeners(Collections.singletonList(listener)); cleaner.setListeners(Collections.singletonList(listener));
} }
@@ -97,7 +101,6 @@ public class ContentStoreCleanerTest extends TestCase
// the content should have disappeared as it is not in the database // the content should have disappeared as it is not in the database
assertFalse("Unprotected content was not deleted", store.exists(contentUrl)); assertFalse("Unprotected content was not deleted", store.exists(contentUrl));
assertTrue("Content listener was not called with deletion", deletedUrls.contains(contentUrl));
} }
public void testProtectedRemoval() throws Exception public void testProtectedRemoval() throws Exception
@@ -116,6 +119,29 @@ public class ContentStoreCleanerTest extends TestCase
assertFalse("Content listener was called with deletion of protected URL", deletedUrls.contains(contentUrl)); assertFalse("Content listener was called with deletion of protected URL", deletedUrls.contains(contentUrl));
} }
public void testConcurrentRemoval() throws Exception
{
int threadCount = 2;
final CountDownLatch endLatch = new CountDownLatch(threadCount);
// Kick off the threads
for (int i = 0; i < threadCount; i++)
{
Thread thread = new Thread()
{
@Override
public void run()
{
cleaner.execute();
// Notify of completion
endLatch.countDown();
}
};
thread.start();
}
// Wait for them all to be done
endLatch.await();
}
private class DummyCleanerListener implements ContentStoreCleanerListener private class DummyCleanerListener implements ContentStoreCleanerListener
{ {
public void beforeDelete(ContentReader reader) throws ContentIOException public void beforeDelete(ContentReader reader) throws ContentIOException

View File

@@ -29,8 +29,6 @@ import java.io.IOException;
import java.util.Calendar; import java.util.Calendar;
import java.util.Date; import java.util.Date;
import java.util.GregorianCalendar; import java.util.GregorianCalendar;
import java.util.HashSet;
import java.util.Set;
import org.alfresco.error.AlfrescoRuntimeException; import org.alfresco.error.AlfrescoRuntimeException;
import org.alfresco.repo.content.AbstractContentStore; import org.alfresco.repo.content.AbstractContentStore;
@@ -398,29 +396,26 @@ public class FileContentStore extends AbstractContentStore
} }
} }
public Set<String> getUrls(Date createdAfter, Date createdBefore) public void getUrls(Date createdAfter, Date createdBefore, ContentUrlHandler handler)
{ {
// recursively get all files within the root // recursively get all files within the root
Set<String> contentUrls = new HashSet<String>(1000); getUrls(rootDirectory, handler, createdAfter, createdBefore);
getUrls(rootDirectory, contentUrls, createdAfter, createdBefore);
// done // done
if (logger.isDebugEnabled()) if (logger.isDebugEnabled())
{ {
logger.debug("Listed all content URLS: \n" + logger.debug("Listed all content URLS: \n" +
" store: " + this + "\n" + " store: " + this);
" count: " + contentUrls.size());
} }
return contentUrls;
} }
/** /**
* @param directory the current directory to get the files from * @param directory the current directory to get the files from
* @param contentUrls the list of current content URLs to add to * @param handler the callback to use for each URL
* @param createdAfter only get URLs for content create after this date * @param createdAfter only get URLs for content create after this date
* @param createdBefore only get URLs for content created before this date * @param createdBefore only get URLs for content created before this date
* @return Returns a list of all files within the given directory and all subdirectories * @return Returns a list of all files within the given directory and all subdirectories
*/ */
private void getUrls(File directory, Set<String> contentUrls, Date createdAfter, Date createdBefore) private void getUrls(File directory, ContentUrlHandler handler, Date createdAfter, Date createdBefore)
{ {
File[] files = directory.listFiles(); File[] files = directory.listFiles();
if (files == null) if (files == null)
@@ -433,7 +428,7 @@ public class FileContentStore extends AbstractContentStore
if (file.isDirectory()) if (file.isDirectory())
{ {
// we have a subdirectory - recurse // we have a subdirectory - recurse
getUrls(file, contentUrls, createdAfter, createdBefore); getUrls(file, handler, createdAfter, createdBefore);
} }
else else
{ {
@@ -451,7 +446,8 @@ public class FileContentStore extends AbstractContentStore
} }
// found a file - create the URL // found a file - create the URL
String contentUrl = makeContentUrl(file); String contentUrl = makeContentUrl(file);
contentUrls.add(contentUrl); // Callback
handler.handle(contentUrl);
} }
} }
} }

View File

@@ -626,7 +626,16 @@ abstract public class AbstractMappingMetadataExtracter implements MetadataExtrac
Map<QName, Serializable> changedProperties = null; Map<QName, Serializable> changedProperties = null;
try try
{ {
Map<String, Serializable> rawMetadata = extractRaw(reader); Map<String, Serializable> rawMetadata = null;
// Check that the content has some meat
if (reader.getSize() > 0 && reader.exists())
{
rawMetadata = extractRaw(reader);
}
else
{
rawMetadata = new HashMap<String, Serializable>(1);
}
// Convert to system properties (standalone) // Convert to system properties (standalone)
Map<QName, Serializable> systemProperties = mapRawToSystem(rawMetadata); Map<QName, Serializable> systemProperties = mapRawToSystem(rawMetadata);
// Convert the properties according to the dictionary types // Convert the properties according to the dictionary types

View File

@@ -28,6 +28,7 @@ import java.io.File;
import java.io.FileNotFoundException; import java.io.FileNotFoundException;
import java.io.Serializable; import java.io.Serializable;
import java.util.HashMap; import java.util.HashMap;
import java.util.List;
import java.util.Map; import java.util.Map;
import junit.framework.TestCase; import junit.framework.TestCase;
@@ -35,12 +36,15 @@ import junit.framework.TestCase;
import org.alfresco.model.ContentModel; import org.alfresco.model.ContentModel;
import org.alfresco.repo.content.MimetypeMap; import org.alfresco.repo.content.MimetypeMap;
import org.alfresco.repo.content.filestore.FileContentReader; import org.alfresco.repo.content.filestore.FileContentReader;
import org.alfresco.repo.content.filestore.FileContentWriter;
import org.alfresco.repo.content.transform.AbstractContentTransformerTest; import org.alfresco.repo.content.transform.AbstractContentTransformerTest;
import org.alfresco.service.cmr.dictionary.DictionaryService; import org.alfresco.service.cmr.dictionary.DictionaryService;
import org.alfresco.service.cmr.repository.ContentReader; import org.alfresco.service.cmr.repository.ContentReader;
import org.alfresco.service.cmr.repository.ContentWriter;
import org.alfresco.service.cmr.repository.datatype.DefaultTypeConverter; import org.alfresco.service.cmr.repository.datatype.DefaultTypeConverter;
import org.alfresco.service.namespace.QName; import org.alfresco.service.namespace.QName;
import org.alfresco.util.ApplicationContextHelper; import org.alfresco.util.ApplicationContextHelper;
import org.alfresco.util.PropertyMap;
import org.alfresco.util.TempFileProvider; import org.alfresco.util.TempFileProvider;
import org.springframework.context.ApplicationContext; import org.springframework.context.ApplicationContext;
@@ -134,4 +138,29 @@ public abstract class AbstractMetadataExtracterTest extends TestCase
QUICK_DESCRIPTION, QUICK_DESCRIPTION,
DefaultTypeConverter.INSTANCE.convert(String.class, properties.get(ContentModel.PROP_DESCRIPTION))); DefaultTypeConverter.INSTANCE.convert(String.class, properties.get(ContentModel.PROP_DESCRIPTION)));
} }
public void testZeroLengthFile() throws Exception
{
MetadataExtracter extractor = getExtracter();
File file = TempFileProvider.createTempFile(getName(), ".bin");
ContentWriter writer = new FileContentWriter(file);
writer.getContentOutputStream().close();
ContentReader reader = writer.getReader();
// Try the zero length file against all supported mimetypes.
// Note: Normally the reader would need to be fetched for each access, but we need to be sure
// that the content is not accessed on the reader AT ALL.
PropertyMap properties = new PropertyMap();
List<String> mimetypes = mimetypeMap.getMimetypes();
for (String mimetype : mimetypes)
{
if (!extractor.isSupported(mimetype))
{
// Not interested
continue;
}
reader.setMimetype(mimetype);
extractor.extract(reader, properties);
assertEquals("There should not be any new properties", 0, properties.size());
}
}
} }

View File

@@ -24,9 +24,8 @@
*/ */
package org.alfresco.repo.content.replication; package org.alfresco.repo.content.replication;
import java.util.Set;
import org.alfresco.repo.content.ContentStore; import org.alfresco.repo.content.ContentStore;
import org.alfresco.repo.content.ContentStore.ContentUrlHandler;
import org.alfresco.repo.node.index.IndexRecovery; import org.alfresco.repo.node.index.IndexRecovery;
import org.alfresco.service.cmr.repository.ContentReader; import org.alfresco.service.cmr.repository.ContentReader;
import org.alfresco.service.cmr.repository.ContentWriter; import org.alfresco.service.cmr.repository.ContentWriter;
@@ -149,23 +148,28 @@ public class ContentStoreReplicator
} }
} }
/**
* Handler that does the actual replication
*
* @author Derek Hulley
* @since 2.0
*/
private class ReplicatingHandler implements ContentUrlHandler
{
public void handle(String contentUrl)
{
replicate(contentUrl);
}
}
/** /**
* Perform a full replication of all source to target URLs. * Perform a full replication of all source to target URLs.
*/ */
private void replicate() private void replicate()
{ {
// get all the URLs from the source ReplicatingHandler handler = new ReplicatingHandler();
Set<String> sourceUrls = sourceStore.getUrls(); // Iterate over all the URLs
// get all the URLs from the target sourceStore.getUrls(handler);
Set<String> targetUrls = targetStore.getUrls();
// remove source URLs that are present in the target
sourceUrls.removeAll(targetUrls);
// ensure that each remaining source URL is present in the target
for (String contentUrl : sourceUrls)
{
replicate(contentUrl);
}
} }
/** /**

View File

@@ -25,6 +25,7 @@
package org.alfresco.repo.content.replication; package org.alfresco.repo.content.replication;
import java.io.File; import java.io.File;
import java.util.HashSet;
import java.util.Set; import java.util.Set;
import junit.framework.TestCase; import junit.framework.TestCase;
@@ -32,6 +33,7 @@ import junit.framework.TestCase;
import org.alfresco.repo.content.AbstractContentStore; import org.alfresco.repo.content.AbstractContentStore;
import org.alfresco.repo.content.ContentContext; import org.alfresco.repo.content.ContentContext;
import org.alfresco.repo.content.ContentStore; import org.alfresco.repo.content.ContentStore;
import org.alfresco.repo.content.ContentStore.ContentUrlHandler;
import org.alfresco.repo.content.filestore.FileContentStore; import org.alfresco.repo.content.filestore.FileContentStore;
import org.alfresco.service.cmr.repository.ContentWriter; import org.alfresco.service.cmr.repository.ContentWriter;
import org.alfresco.util.GUID; import org.alfresco.util.GUID;
@@ -106,6 +108,21 @@ public class ContentStoreReplicatorTest extends TestCase
targetStore.exists(writer.getContentUrl())); targetStore.exists(writer.getContentUrl()));
} }
/**
* Handler that merely records the URL
*
* @author Derek Hulley
* @since 2.0
*/
private class UrlRecorder implements ContentUrlHandler
{
public Set<String> urls = new HashSet<String>(1027);
public void handle(String contentUrl)
{
urls.add(contentUrl);
}
}
/** /**
* Adds content to the source while the replicator is going as fast as possible. * Adds content to the source while the replicator is going as fast as possible.
* Just to make it more interesting, the content is sometimes put in the target * Just to make it more interesting, the content is sometimes put in the target
@@ -150,11 +167,13 @@ public class ContentStoreReplicatorTest extends TestCase
} }
// check that we have an exact match of URLs // check that we have an exact match of URLs
Set<String> sourceUrls = sourceStore.getUrls(); UrlRecorder sourceUrls = new UrlRecorder();
Set<String> targetUrls = targetStore.getUrls(); UrlRecorder targetUrls = new UrlRecorder();
sourceStore.getUrls(sourceUrls);
targetStore.getUrls(targetUrls);
sourceUrls.containsAll(targetUrls); sourceUrls.urls.containsAll(targetUrls.urls);
targetUrls.contains(sourceUrls); targetUrls.urls.contains(sourceUrls.urls);
} }
/** /**

View File

@@ -25,9 +25,7 @@
package org.alfresco.repo.content.replication; package org.alfresco.repo.content.replication;
import java.util.Date; import java.util.Date;
import java.util.HashSet;
import java.util.List; import java.util.List;
import java.util.Set;
import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.ThreadPoolExecutor;
import java.util.concurrent.locks.Lock; import java.util.concurrent.locks.Lock;
import java.util.concurrent.locks.ReadWriteLock; import java.util.concurrent.locks.ReadWriteLock;
@@ -370,32 +368,26 @@ public class ReplicatingContentStore extends AbstractContentStore
} }
/** /**
* @return Returns the results as given by the primary store, and if inbound * Iterates over results as given by the primary store and all secondary stores. It is up to the handler to eliminate
* replication is active, merges the URLs from the secondary stores. * duplicates that will occur between the primary and secondary stores.
*/ */
public Set<String> getUrls(Date createdAfter, Date createdBefore) throws ContentIOException public void getUrls(Date createdAfter, Date createdBefore, ContentUrlHandler handler) throws ContentIOException
{ {
Set<String> urls = new HashSet<String>(1024);
// add in URLs from primary store // add in URLs from primary store
Set<String> primaryUrls = primaryStore.getUrls(createdAfter, createdBefore); primaryStore.getUrls(createdAfter, createdBefore, handler);
urls.addAll(primaryUrls);
// add in URLs from secondary stores (they are visible for reads) // add in URLs from secondary stores (they are visible for reads)
for (ContentStore secondaryStore : secondaryStores) for (ContentStore secondaryStore : secondaryStores)
{ {
Set<String> secondaryUrls = secondaryStore.getUrls(createdAfter, createdBefore); secondaryStore.getUrls(createdAfter, createdBefore, handler);
// merge them
urls.addAll(secondaryUrls);
} }
// done // done
if (logger.isDebugEnabled()) if (logger.isDebugEnabled())
{ {
logger.debug("Found " + urls.size() + " URLs, of which " + primaryUrls.size() + " are primary: \n" + logger.debug("Iterated over content URLs: \n" +
" created after: " + createdAfter + "\n" + " created after: " + createdAfter + "\n" +
" created before: " + createdBefore); " created before: " + createdBefore);
} }
return urls;
} }
/** /**

View File

@@ -26,6 +26,7 @@ package org.alfresco.repo.content.replication;
import java.io.File; import java.io.File;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.HashSet;
import java.util.List; import java.util.List;
import java.util.Set; import java.util.Set;
import java.util.concurrent.SynchronousQueue; import java.util.concurrent.SynchronousQueue;
@@ -35,6 +36,7 @@ import java.util.concurrent.TimeUnit;
import org.alfresco.repo.content.AbstractWritableContentStoreTest; import org.alfresco.repo.content.AbstractWritableContentStoreTest;
import org.alfresco.repo.content.ContentContext; import org.alfresco.repo.content.ContentContext;
import org.alfresco.repo.content.ContentStore; import org.alfresco.repo.content.ContentStore;
import org.alfresco.repo.content.ContentStore.ContentUrlHandler;
import org.alfresco.repo.content.filestore.FileContentStore; import org.alfresco.repo.content.filestore.FileContentStore;
import org.alfresco.service.cmr.repository.ContentIOException; import org.alfresco.service.cmr.repository.ContentIOException;
import org.alfresco.service.cmr.repository.ContentReader; import org.alfresco.service.cmr.repository.ContentReader;
@@ -135,7 +137,15 @@ public class ReplicatingContentStoreTest extends AbstractWritableContentStoreTes
// check that the URL is present for each of the stores // check that the URL is present for each of the stores
for (ContentStore store : secondaryStores) for (ContentStore store : secondaryStores)
{ {
Set<String> urls = store.getUrls(); final Set<String> urls = new HashSet<String>(1027);
ContentUrlHandler handler = new ContentUrlHandler()
{
public void handle(String contentUrl)
{
urls.add(contentUrl);
}
};
store.getUrls(handler);
assertTrue("URL of new content not present in store", urls.contains(contentUrl) == mustExist); assertTrue("URL of new content not present in store", urls.contains(contentUrl) == mustExist);
} }
} }

View File

@@ -0,0 +1,49 @@
/*
* Copyright (C) 2005-2007 Alfresco Software Limited.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version 2
* of the License, or (at your option) any later version.
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
* As a special exception to the terms and conditions of version 2.0 of
* the GPL, you may redistribute this Program in connection with Free/Libre
* and Open Source Software ("FLOSS") applications as described in Alfresco's
* FLOSS exception. You should have recieved a copy of the text describing
* the FLOSS exception, and it is also available here:
* http://www.alfresco.com/legal/licensing
*/
package org.alfresco.repo.domain;
/**
* Interface for persistent <b>Content URL</b> objects.
* <p>
* Instances represent physically stored content.
*
* @author Derek Hulley
* @since 2.0
*/
public interface ContentUrl
{
/**
* @return Returns the auto-generated ID
*/
Long getId();
String getContentUrl();
void setContentUrl(String contentUrl);
//
// boolean isOrphaned();
//
// void setOrphaned(boolean orphaned);
}

View File

@@ -0,0 +1,73 @@
/*
* Copyright (C) 2005-2007 Alfresco Software Limited.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version 2
* of the License, or (at your option) any later version.
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
* As a special exception to the terms and conditions of version 2.0 of
* the GPL, you may redistribute this Program in connection with Free/Libre
* and Open Source Software ("FLOSS") applications as described in Alfresco's
* FLOSS exception. You should have recieved a copy of the text describing
* the FLOSS exception, and it is also available here:
* http://www.alfresco.com/legal/licensing */
package org.alfresco.repo.domain;
import java.util.Set;
/**
* Abstraction for manipulating <b>Content URL</b> entities.
*
* @author Derek Hulley
* @since 2.0
*/
public interface ContentUrlDAO
{
/**
* Create a new <b>Content URL</b> or get an existing instance.
*/
ContentUrl createContentUrl(String contentUrl);
/**
* Enumerate all the available <b>Content URLs</b>, calling back to the given handler.
*
* @param handler the component that will be called with each URL
*/
void getAllContentUrls(ContentUrlHandler handler);
/**
* Delete the <b>Content URL</b>.
*/
void deleteContentUrl(String contentUrl);
/**
* Delete a set of <b>Content URL</b>.
*/
void deleteContentUrls(Set<String> contentUrls);
/**
* Delete all <b>Content URL</b> entities.
*/
void deleteAllContentUrls();
/**
* A callback interface to handle <b>Content URLS<b> produced by iteration.
*
* @author Derek Hulley
* @since 2.0
*/
public interface ContentUrlHandler
{
void handle(String contentUrl);
};
}

View File

@@ -0,0 +1,212 @@
package org.alfresco.repo.domain;
import java.util.HashSet;
import java.util.Set;
import javax.transaction.UserTransaction;
import junit.framework.TestCase;
import org.alfresco.repo.content.filestore.FileContentStore;
import org.alfresco.repo.security.authentication.AuthenticationUtil;
import org.alfresco.repo.security.authentication.AuthenticationUtil.RunAsWork;
import org.alfresco.service.ServiceRegistry;
import org.alfresco.service.cmr.repository.ContentService;
import org.alfresco.service.transaction.TransactionService;
import org.alfresco.util.ApplicationContextHelper;
import org.springframework.context.ApplicationContext;
/**
* @see org.alfresco.repo.domain.ContentUrlDAO
*
* @author Derek Hulley
* @since 2.1
*/
public class ContentUrlDAOTest extends TestCase
{
private static ApplicationContext ctx = ApplicationContextHelper.getApplicationContext();
private ContentUrlDAO dao;
private TransactionService transactionService;
private ContentService contentService;
@Override
protected void setUp() throws Exception
{
ServiceRegistry serviceRegistry = (ServiceRegistry) ctx.getBean(ServiceRegistry.SERVICE_REGISTRY);
dao = (ContentUrlDAO) ctx.getBean("contentUrlDAO");
contentService = serviceRegistry.getContentService();
transactionService = serviceRegistry.getTransactionService();
}
@Override
protected void tearDown() throws Exception
{
}
public void testCreateContentUrl() throws Throwable
{
UserTransaction txn = transactionService.getUserTransaction();
try
{
txn.begin();
RunAsWork<String> getTempWriterWork = new RunAsWork<String>()
{
public String doWork() throws Exception
{
return contentService.getTempWriter().getContentUrl();
}
};
String contentUrl = AuthenticationUtil.runAs(getTempWriterWork, AuthenticationUtil.SYSTEM_USER_NAME);
// Make sure that it can be written in duplicate
ContentUrl entity1 = dao.createContentUrl(contentUrl);
ContentUrl entity2 = dao.createContentUrl(contentUrl);
assertNotSame("Assigned IDs must be new", entity1.getId(), entity2.getId());
txn.commit();
}
catch (Throwable e)
{
try { txn.rollback(); } catch (Throwable ee) {}
throw e;
}
}
private Set<String> makeUrls(int count) throws Throwable
{
final Set<String> urls = new HashSet<String>(count);
for (int i = 0; i < count; i++)
{
String contentUrl = String.format("%s%s/%04d", FileContentStore.STORE_PROTOCOL, getName(), i);
dao.createContentUrl(contentUrl);
urls.add(contentUrl);
}
return urls;
}
public void testGetAllContentUrls() throws Throwable
{
UserTransaction txn = transactionService.getUserTransaction();
try
{
txn.begin();
final Set<String> urls = makeUrls(1000);
// Now iterate over them in the same transaction
ContentUrlDAO.ContentUrlHandler handler = new ContentUrlDAO.ContentUrlHandler()
{
public void handle(String contentUrl)
{
urls.remove(contentUrl);
}
};
dao.getAllContentUrls(handler);
assertEquals("Not all content URLs were enumerated", 0, urls.size());
txn.commit();
}
catch (Throwable e)
{
try { txn.rollback(); } catch (Throwable ee) {}
throw e;
}
}
public void testDeleteContentUrl() throws Throwable
{
UserTransaction txn = transactionService.getUserTransaction();
try
{
txn.begin();
final Set<String> urls = makeUrls(1000);
// Delete them
for (String url : urls)
{
dao.deleteContentUrl(url);
}
// Now iterate over them in the same transaction
ContentUrlDAO.ContentUrlHandler handler = new ContentUrlDAO.ContentUrlHandler()
{
public void handle(String contentUrl)
{
urls.remove(contentUrl);
}
};
dao.getAllContentUrls(handler);
// All the URLs previously deleted will not have been removed from the Set
assertEquals("Specific content URLs were not deleted", 1000, urls.size());
txn.commit();
}
catch (Throwable e)
{
try { txn.rollback(); } catch (Throwable ee) {}
throw e;
}
}
public void testDeleteContentUrls() throws Throwable
{
UserTransaction txn = transactionService.getUserTransaction();
try
{
txn.begin();
final Set<String> urls = makeUrls(1000);
// Delete them
dao.deleteContentUrls(urls);
// Now iterate over them in the same transaction
ContentUrlDAO.ContentUrlHandler handler = new ContentUrlDAO.ContentUrlHandler()
{
public void handle(String contentUrl)
{
urls.remove(contentUrl);
}
};
dao.getAllContentUrls(handler);
// All the URLs previously deleted will not have been removed from the Set
assertEquals("Specific content URLs were not deleted", 1000, urls.size());
txn.commit();
}
catch (Throwable e)
{
try { txn.rollback(); } catch (Throwable ee) {}
throw e;
}
}
public void testDeleteAllContentUrls() throws Throwable
{
UserTransaction txn = transactionService.getUserTransaction();
try
{
txn.begin();
makeUrls(1000);
// Delete them
dao.deleteAllContentUrls();
// Check that there are none left
// Now iterate over them in the same transaction
ContentUrlDAO.ContentUrlHandler handler = new ContentUrlDAO.ContentUrlHandler()
{
public void handle(String contentUrl)
{
fail("There should not be any URLs remaining.");
}
};
dao.getAllContentUrls(handler);
txn.commit();
}
catch (Throwable e)
{
try { txn.rollback(); } catch (Throwable ee) {}
throw e;
}
}
}

View File

@@ -0,0 +1,56 @@
<?xml version='1.0' encoding='UTF-8'?>
<!DOCTYPE hibernate-mapping PUBLIC
'-//Hibernate/Hibernate Mapping DTD 3.0//EN'
'http://hibernate.sourceforge.net/hibernate-mapping-3.0.dtd'>
<hibernate-mapping>
<class
name="org.alfresco.repo.domain.hibernate.ContentUrlImpl"
proxy="org.alfresco.repo.domain.ContentUrl"
table="alf_content_url"
dynamic-update="false"
dynamic-insert="false"
select-before-update="false"
optimistic-lock="none" >
<!-- auto-generated ID -->
<id name="id" column="id" type="long" >
<generator class="increment" />
</id>
<property name="contentUrl" column="content_url" type="string" length="240" not-null="true" index="idx_alf_con_urls" />
<!--
<property name="isOrphaned" column="is_orphaned" type="boolean" not-null="true" />
-->
</class>
<query name="contentUrl.GetAll">
select
entity.contentUrl
from
org.alfresco.repo.domain.hibernate.ContentUrlImpl entity
</query>
<query name="contentUrl.DeleteInList">
delete
from
org.alfresco.repo.domain.hibernate.ContentUrlImpl entity
where
entity.contentUrl in (:contentUrls)
</query>
<query name="contentUrl.DeleteByUrl">
delete
from
org.alfresco.repo.domain.hibernate.ContentUrlImpl entity
where
entity.contentUrl = :contentUrl
</query>
<query name="contentUrl.DeleteAll">
delete
from
org.alfresco.repo.domain.hibernate.ContentUrlImpl entity
</query>
</hibernate-mapping>

View File

@@ -0,0 +1,86 @@
/*
* Copyright (C) 2005-2007 Alfresco Software Limited.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version 2
* of the License, or (at your option) any later version.
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
* As a special exception to the terms and conditions of version 2.0 of
* the GPL, you may redistribute this Program in connection with Free/Libre
* and Open Source Software ("FLOSS") applications as described in Alfresco's
* FLOSS exception. You should have recieved a copy of the text describing
* the FLOSS exception, and it is also available here:
* http://www.alfresco.com/legal/licensing
*/
package org.alfresco.repo.domain.hibernate;
import java.io.Serializable;
import org.alfresco.repo.domain.ContentUrl;
/**
* Bean containing all the persistence data representing a <b>Content Url</b>.
* <p>
* This implementation of the {@link org.alfresco.repo.domain.Node Node} interface is
* Hibernate specific.
*
* @author Derek Hulley
* @since 2.0
*/
public class ContentUrlImpl extends LifecycleAdapter implements ContentUrl, Serializable
{
private static final long serialVersionUID = -7368859912728834288L;
private Long id;
private String contentUrl;
// private boolean isOrphaned;
public ContentUrlImpl()
{
// isOrphaned = false;
}
public Long getId()
{
return id;
}
/**
* For Hibernate Use
*/
@SuppressWarnings("unused")
private void setId(Long id)
{
this.id = id;
}
public String getContentUrl()
{
return contentUrl;
}
public void setContentUrl(String contentUrl)
{
this.contentUrl = contentUrl;
}
//
// public boolean isOrphaned()
// {
// return isOrphaned;
// }
//
// public void setOrphaned(boolean isOrphaned)
// {
// this.isOrphaned = isOrphaned;
// }
}

View File

@@ -0,0 +1,123 @@
package org.alfresco.repo.domain.hibernate;
import java.util.Set;
import org.alfresco.repo.domain.ContentUrl;
import org.alfresco.repo.domain.ContentUrlDAO;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.hibernate.CacheMode;
import org.hibernate.Query;
import org.hibernate.ScrollMode;
import org.hibernate.ScrollableResults;
import org.hibernate.Session;
import org.hibernate.type.TypeFactory;
import org.springframework.orm.hibernate3.HibernateCallback;
import org.springframework.orm.hibernate3.support.HibernateDaoSupport;
/**
* Hibernate-specific implementation of the DAO layer for <b>Content URLs</b>.
*
* @author Derek Hulley
* @since 2.0
*/
public class HibernateContentUrlDAOImpl extends HibernateDaoSupport implements ContentUrlDAO
{
private static final String QUERY_GET_ALL = "contentUrl.GetAll";
private static final String UPDATE_DELETE_BY_URL = "contentUrl.DeleteByUrl";
private static final String UPDATE_DELETE_IN_LIST = "contentUrl.DeleteInList";
private static final String UPDATE_DELETE_ALL = "contentUrl.DeleteAll";
private static Log logger = LogFactory.getLog(HibernateContentUrlDAOImpl.class);
public ContentUrl createContentUrl(String contentUrl)
{
ContentUrl entity = new ContentUrlImpl();
entity.setContentUrl(contentUrl);
getSession().save(entity);
return entity;
}
public void getAllContentUrls(final ContentUrlHandler handler)
{
HibernateCallback callback = new HibernateCallback()
{
public Object doInHibernate(Session session)
{
Query query = session
.getNamedQuery(HibernateContentUrlDAOImpl.QUERY_GET_ALL)
.setCacheMode(CacheMode.IGNORE);
return query.scroll(ScrollMode.FORWARD_ONLY);
}
};
ScrollableResults results = (ScrollableResults) getHibernateTemplate().execute(callback);
while (results.next())
{
String contentUrl = results.getText(0);
handler.handle(contentUrl);
}
}
public void deleteContentUrl(final String contentUrl)
{
HibernateCallback callback = new HibernateCallback()
{
public Object doInHibernate(Session session)
{
session.flush();
Query query = session
.getNamedQuery(HibernateContentUrlDAOImpl.UPDATE_DELETE_BY_URL)
.setCacheMode(CacheMode.IGNORE)
.setString("contentUrl", contentUrl);
return (Integer) query.executeUpdate();
}
};
Integer deletedCount = (Integer) getHibernateTemplate().execute(callback);
int entityCount = getSession().getStatistics().getEntityCount();
if (logger.isDebugEnabled())
{
logger.debug("Deleted " + deletedCount + " ContentUrl entities.");
}
}
public void deleteContentUrls(final Set<String> contentUrls)
{
HibernateCallback callback = new HibernateCallback()
{
public Object doInHibernate(Session session)
{
session.flush();
Query query = session
.getNamedQuery(HibernateContentUrlDAOImpl.UPDATE_DELETE_IN_LIST)
.setCacheMode(CacheMode.IGNORE)
.setParameterList("contentUrls", contentUrls, TypeFactory.basic("string"));
return (Integer) query.executeUpdate();
}
};
Integer deletedCount = (Integer) getHibernateTemplate().execute(callback);
if (logger.isDebugEnabled())
{
logger.debug("Deleted " + deletedCount + " ContentUrl entities.");
}
}
public void deleteAllContentUrls()
{
HibernateCallback callback = new HibernateCallback()
{
public Object doInHibernate(Session session)
{
session.flush();
Query query = session
.getNamedQuery(HibernateContentUrlDAOImpl.UPDATE_DELETE_ALL)
.setCacheMode(CacheMode.IGNORE);
return (Integer) query.executeUpdate();
}
};
Integer deletedCount = (Integer) getHibernateTemplate().execute(callback);
if (logger.isDebugEnabled())
{
logger.debug("Deleted " + deletedCount + " ContentUrl entities.");
}
}
}

View File

@@ -39,6 +39,7 @@ import org.alfresco.repo.domain.ChildAssoc;
import org.alfresco.repo.domain.Node; import org.alfresco.repo.domain.Node;
import org.alfresco.repo.domain.NodeStatus; import org.alfresco.repo.domain.NodeStatus;
import org.alfresco.repo.node.BaseNodeServiceTest; import org.alfresco.repo.node.BaseNodeServiceTest;
import org.alfresco.repo.node.db.NodeDaoService.NodePropertyHandler;
import org.alfresco.repo.transaction.AlfrescoTransactionSupport; import org.alfresco.repo.transaction.AlfrescoTransactionSupport;
import org.alfresco.repo.transaction.RetryingTransactionHelper.RetryingTransactionCallback; import org.alfresco.repo.transaction.RetryingTransactionHelper.RetryingTransactionCallback;
import org.alfresco.service.cmr.dictionary.DataTypeDefinition; import org.alfresco.service.cmr.dictionary.DataTypeDefinition;
@@ -297,7 +298,15 @@ public class DbNodeServiceImplTest extends BaseNodeServiceTest
collection); collection);
// get a list of all content values // get a list of all content values
List<Serializable> allContentDatas = nodeDaoService.getPropertyValuesByActualType(contentDataType); final List<Serializable> allContentDatas = new ArrayList<Serializable>(500);
NodePropertyHandler handler = new NodePropertyHandler()
{
public void handle(Node node, Serializable value)
{
allContentDatas.add(value);
}
};
nodeDaoService.getPropertyValuesByActualType(contentDataType, handler);
assertTrue("At least two instances expected", allContentDatas.size() >= 2); assertTrue("At least two instances expected", allContentDatas.size() >= 2);
assertTrue("Single content data not present in results", assertTrue("Single content data not present in results",
allContentDatas.contains(contentDataSingle)); allContentDatas.contains(contentDataSingle));

View File

@@ -277,12 +277,20 @@ public interface NodeDaoService
public void deleteNodeAssoc(NodeAssoc assoc); public void deleteNodeAssoc(NodeAssoc assoc);
/** /**
* Fetch all property values for the given type definition. This will also dig out values that * Iterate over all property values for the given type definition. This will also dig out values that
* were persisted as type <b>d:any</b>. * were persisted as type <b>d:any</b>.
* *
* @param actualDataTypeDefinition the persisted type to retrieve
* @param handler the callback to use while iterating over the URLs
* @return Returns the values for the given type definition * @return Returns the values for the given type definition
*/ */
public List<Serializable> getPropertyValuesByActualType(DataTypeDefinition actualDataTypeDefinition); public void getPropertyValuesByActualType(DataTypeDefinition actualDataTypeDefinition, NodePropertyHandler handler);
/**
* Get properties with the given type and string value.
* TODO: Refactor as in getPropertyValuesByActualType
*/
public Collection<Node> getNodesWithPropertyStringValueForStore(StoreRef storeRef, QName propQName, String propStringValue);
/** /**
* @return Returns the total number of nodes in the ADM repository * @return Returns the total number of nodes in the ADM repository
@@ -293,7 +301,16 @@ public interface NodeDaoService
*/ */
public int getNodeCount(final StoreRef storeRef); public int getNodeCount(final StoreRef storeRef);
public Collection<Node> getNodesWithPropertyStringValueForStore(final StoreRef storeRef, final QName propQName, final String propStringValue); /**
* Iterface to handle callbacks when iterating over properties
*
* @author Derek Hulley
* @since 2.0
*/
public interface NodePropertyHandler
{
void handle(Node node, Serializable value);
}
public Transaction getTxnById(long txnId); public Transaction getTxnById(long txnId);
/** /**

View File

@@ -1376,7 +1376,7 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
getHibernateTemplate().delete(assoc); getHibernateTemplate().delete(assoc);
} }
public List<Serializable> getPropertyValuesByActualType(DataTypeDefinition actualDataTypeDefinition) public void getPropertyValuesByActualType(DataTypeDefinition actualDataTypeDefinition, NodePropertyHandler handler)
{ {
// get the in-database string representation of the actual type // get the in-database string representation of the actual type
QName typeQName = actualDataTypeDefinition.getName(); QName typeQName = actualDataTypeDefinition.getName();
@@ -1393,7 +1393,6 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
}; };
ScrollableResults results = (ScrollableResults) getHibernateTemplate().execute(callback); ScrollableResults results = (ScrollableResults) getHibernateTemplate().execute(callback);
// Loop through, extracting content URLs // Loop through, extracting content URLs
List<Serializable> convertedValues = new ArrayList<Serializable>(1000);
TypeConverter converter = DefaultTypeConverter.INSTANCE; TypeConverter converter = DefaultTypeConverter.INSTANCE;
int unflushedCount = 0; int unflushedCount = 0;
while(results.next()) while(results.next())
@@ -1418,16 +1417,19 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
{ {
continue; continue;
} }
Serializable convertedValue = null;
try try
{ {
Serializable convertedValue = (Serializable) converter.convert(actualDataTypeDefinition, value); convertedValue = (Serializable) converter.convert(actualDataTypeDefinition, value);
// it converted, so add it
convertedValues.add(convertedValue);
} }
catch (Throwable e) catch (Throwable e)
{ {
// The value can't be converted - forget it // The value can't be converted - forget it
} }
if (convertedValue != null)
{
handler.handle(node, convertedValue);
}
} }
} }
unflushedCount++; unflushedCount++;
@@ -1438,7 +1440,6 @@ public class HibernateNodeDaoServiceImpl extends HibernateDaoSupport implements
unflushedCount = 0; unflushedCount = 0;
} }
} }
return convertedValues;
} }
/** /**