Dave Ward ee8081b663 Merged V4.1-BUG-FIX to HEAD
41035: Fix for ALF-15225 - qt.length not performing as expected in search.lib.js
   41047: RUSSIAN: Translation updates based on EN r40961
   41049: GERMAN: Translation updates based on EN r41033. Fixes ALF-15749 and ALF-15720.
   41050: SPANISH: Translation updates based on EN r41033. Fixes ALF-15749 and ALF-15720.
   41051: FRENCH: Translation updates based on EN r41033. Fixes ALF-15749 and ALF-15720.
   41052: ITALIAN: Translation updates based on EN r41033. Fixes ALF-15749 and ALF-15720.
   41053: JAPANESE: Translation updates based on EN r41033. Fixes ALF-15749 and ALF-15720.
   41054: DUTCH: Translation updates based on EN r41033. Fixes ALF-15749 and ALF-15720.
   41055: RUSSIAN: Translation updates based on EN r41033. Fixes ALF-15749 and ALF-15720.
   41061: ALF-11214 - IMAP subsystem is not successfully restarted after incorrect modification of IMAP properties via Admin Console
   41063: RUSSIAN: Translation updates based on EN r41033 (encoding fixes)
   41064: CHINESE: Translation updates based on EN r41033
   41073: ALF-15760: Merged V4.0 to V4.1-BUG-FIX (another lost meta-inf revision)
      34416: ALF-12992: Updated weblogic DD for SOLR
   41074: ALF-15419 / ALF-14438: Merged V3.4-BUG-FIX to V4.1-BUG-FIX (PARTIAL)
      37373: Merged PATCHES/V3.4.6 to V3.4-BUG-FIX
         36821: ALF-13827 / ALF-14402: Make replicated caches recover from temporary comms failures by flushing when a change in peers is detected
         - We do not flush caches who replicate via copy (e.g. tickets cache) as these may not be recoverable
         37122: ALF-13919 / ALF-14403: Merged DEV to PATCHES/V3.4.6
         - Rework of Dmitry's implementation
         - Uses dynamic HQL query to retrieve JBPM workflow instances by specified query criteria
         - WorkflowInstancesGet web script no longer has to iterate over every workflow instance in the database!
         - DB index added to enable efficient querying by string variable
         - Hibernate tastic!
         37188: ALF-13919 / ALF-14403: Worked around HQL polymorphism issues by using explicit variable subclass names in from clause
         37204: ALF-13919 / ALF-14403: Fix to date range handling by Dmitry
   41077: ALF-10730: Fixed object finder drop-down tree to show parent icon type when icon type not available
   41089: ALF-13998: 'No items' error is highlighted in red, even that is not sever error.
   41109: Part fix for ALF-11297 Disabled test: test-system-build-test has been removed from the test target dependency list
   41118: Fixes: ALF-15765 and related issue ALF-15476: 
      - Corrects merge of r34405 and r40159
      - Adds people api to cloud proxy config
      - Cloud/Global Folder picker sites list now universally shows only those sites the user is a member of.
   41119: ALF-15419 / ALF-14438: Merged DEV to V4.1-BUG-FIX
      41117: ALF-15419 : CLONE Remove JBPM indexes present from upgrades
         The schema reference files were updated to contain JBPM_VARIABLEINSTANCE.IDX_VARINST_STRVAL index.
   41130: ALF-15590: FSTR transfer of custom content type with parent cm:content fails
   - Fix by Dmitry
   41131: ALF-15714: Bitrock Uninstaller: Uninstaller doesn't remove all files and folders in *nix
   - Fix provided by Bitrock
   41132: Fix for ALF-14388 - Edit Online option is not supported for '.docm', 'dotm', '.xlsm' files
    - corrected mimetypes to be lower-case
   41150: ALF-13287 Added the handling of UTC represented by "Z" within comparator
   41154: Check for authentication errors when validating a ticket, if an error occurs re-authenticate. ALF-15394
   41155: ALF-15569: User with '@' symbol in username cannot invite
   - note: since Ent 4.0.2(+) by default MT is pre-configured but not enabled
   41173: Merged V4.1 to V4.1-BUG-FIX
      41121: Merged BRANCHES/DEV/FEATURES/CLOUD1_CLOUDSYNC to BRANCHES/V4.1:
         41003: CloudSync: ALF-15734 - force unsync (of last SSMN) on target causes repeating pull errors to appear in both logs
         41026: CloudSync: ALF-15734 - force unsync (of last SSMN) on target causes repeating pull errors to appear in both logs
         41039: CloudSync: ALF-15734 - force unsync (of last SSMN) on target causes repeating pull errors to appear in both logs
         41086: CloudSync: ALF-15734 - force unsync (of last SSMN) on target causes repeating pull errors to appear in both logs
      41123: Merged BRANCHES/DEV/FEATURES/CLOUD1_CLOUDSYNC to BRANCHES/V4.1:
         41115: CloudSync: ALF-15734 - force unsync (of last SSMN) on target causes repeating pull errors to appear in both logs
   41176: Attempt to debug unit test failure
   41181: Store leak in AVMServiceTest.test_ETWOTWO_570() causing unit test failure
   41184: ALF-15610: Copy Thai analyzer settings to its many SOLR locations
   41194: ALF-11297: re-enable system build tests
   41195: ALF-11297 ALF-15807: update activities system build tests after correction of ALF-4832
   41201: Fix for ALF-15767 Group query using cm:authorityName
   41202: Additional unit tests related to ALF-15731  TYPE:"..." queries no longer work for Lucene on 4.X
   41203: Part 1 for ALF-15811 SOLR query increases DocBitSet inefficiently
   - check it makes any difference
   41204: Merged BRANCHES/DEV/BELARUS/V4.1-BUG-FIX-2012_08_15 to BRANCHES/DEV/V4.1-BUG-FIX: (note: merging as-is ... refactor + unit test fix to follow in next commit)
      40926: ALF-12586: Admin Console shows usage as zero - if user is deleted and then re-created (eg. re-synchronized via LDAP or manually)
      40974: ALF-12586: Admin Console shows usage as zero - if user is deleted and then re-created (eg. re-synchronized via LDAP or manually)
   41205: ALF-12586: Admin Console shows usage as zero - if user is deleted and then re-created (eg. re-synchronized via LDAP or manually)
   - review and refactor the proposed/merged fix (see previous commit)
   - fix unit test so that it accounts for previous content (as per the original issue)
   - also: add missing test to suite and fix that test to work with the new fix (ie. cleanup previous content, else need to account for it)
   41210: Part 2:  ALF-14861 SOLR to scale for non-admin users in 100k sites and a subgroup of each of 1000 independent groupings with 1000 subgroups
   - do not expand authorities for thoses with the ADMINISTRATOR_ROLE as they can read all anyway
   41216: ALF-11297: system build tests need a database cleanup before running
   41222: ALF-15740, ALF-14744: Update rule firing broken for content created in Explorer
   - Old code lurking around that used to use the inline editable aspect to detect events handled by the CreateNodeRuleTrigger was removed from OnContentUpdateRuleTrigger and replaced with a check for ASPECT_NO_CONTENT
   41223: Added missing swf.languagedir setting to enterprise alfresco-global.properties
   41230: GERMAN: Translation updates based on EN rev41099.
   41232: SPANISH: Translation updates based on EN rev41099.
   41233: FRENCH: Translation updates based on EN rev41099.
   41234: ITALIAN: Translation updates based on EN rev41099.
   41235: JAPANESE: Translation updates based on EN rev41099.
   41236: DUTCH: Translation updates based on EN rev41099.
   41237: RUSSIAN: Translation updates based on EN rev41099.
   41239: CHINESE: Translation updates based on EN rev41099.
   41254: ALF-15628: Avoid edit online (SPP, WRITE_LOCK) clashing with edit offline (CheckOutCheckInService, READ_ONLY_LOCK)
   - Rationalization of work by Alex Malinovsky
   - WebDAVMethod.checkNode() now properly checks whether nodes without WebDAV lock info are writeable
   - CheckOutCheckInService won't allow checkout of a node with an existing WRITE_LOCK by the same user - they must unlock first
   - Propagation of correct status codes
   41264: ALF-15628: Fix CheckOutCheckInService test failures
   41265: ALF-15699: Reverse merged the following, thus downgrading us back to swftools 0.9.1
      40208: ALF-12831: Upgrade to swftools 0.9.2
   41266: Rush'n in some translation updates from Gloria
   41267: ALF-15628: Fix compilation problem
   41269: Merged V3.4-BUG-FIX to V4.1-BUG-FIX (RECORD ONLY)
      41224: ALF-14856: Merged V4.1-BUG-FIX to V3.4-BUG-FIX
      41268: ALF-15459: Merged PATCHES/V4.0.2 to V3.4-BUG-FIX
         Merged V4.1-BUG-FIX to V3.4-BUG-FIX
   41274: ALF-15608: Merged V3.4-BUG-FIX to V4.1-BUG-FIX
      41272: ALF-15567: Allow links to be followed through WebDAV on port 80 using basic auth on XP
   41277: ALF-12586: Admin Console shows usage as zero - if user is deleted and then re-created (eg. re-synchronized via LDAP or manually) 
   - fix test fallout after merge/fix (note: failed for PostgreSQL but not MySQL - although fix was not DB-specific)
   41278: ALF-15840 Error logged when "No thumbnail present in file" even though this is normal 
   41284: ALF-14875: Serialize direct permissions after inherited permissions so that they take precedence in any lookups in permission dialogs
   41290: JAPANESE: Translation updates based on EN r41099 Fixes: ALF-14565
   41296: ALF-15251	CIFS: Checked out document is not marked as locked in CIFS
   41299: ALF-15714: Bitrock Uninstaller: Uninstaller doesn't remove all files and folders in *nix
   - Additional fix provided by Bitrock
   41303: Fix for ALF-15799 Under high concurrency load balanced Solr throws an Antlr related NPE
   - do not skip IO Exceptions
   41306: Incremented version revision for 4.1.2
   41309: ALF-15827: Added FORMACTION, FORMMETHOD and ACTION  HTML attributes to grey list to close security hole (updated Surf libs r1136)
   41318: ALF-15857: Lucene FTS indexer opens streams to all documents to be indexed in a transaction simultaneously
   - Now stream opening is delayed until the point where the document is being written to the index
   41322: Fix for ALF-15858 SOLR ACL tracking can stall or miss acls during tracking
   41323: Chemistry client java to create test data for ALF-15858, ALF-15782, CLOUD-596, ALF-15753 etc
   41326: ALF-15234: IE 8 or IE9 Download .pps as .ppt 
      -Switched the order of the filename headers to better support non-conforming browsers (rfc 5987).
   41330: ALF-14875: Reversed r41284 because it didn't solve the problem in the UI. Kev reviewing.
   41332: ALF-10688: Can't deactivate an account when alfrescoNtlm follows another authentication subsystem in the authentication chain
   - Now, if a account is known to be 'mutable' then the enabled flag is read from the mutable authentication service
   41337: Merged V3.4-BUG-FIX (3.4.11) to V4.1-BUG-FIX (4.1.2)
      41336: TransformerDebug: Use debug rather than trace when there are transformers but they are all unavailable.
   41339: ALF-15840 Error logged when "No thumbnail present in file" even though this is normal
      - Found some more cases where this is logged as an ERROR
   41342: ALF-11087 (Missing icon file: components\images\filetypes\generic-tag-32.png)
   41344: ALF-15863 (* search values): Merged HEAD to V4.1-BUG-FIX (4.1.2)
      40849: ALF-12839 "Share - Inconsistency in adding a user or a group into a group" part 2
      - Making the users console stop "*" searches, just like the groups console when the min search length is set to larger than 0.
   41346: ALF-15237 - REST API Group children lists username for fullName and displayName
   41350: Merge V3.4-BUG-FIX to V4.1-BUG-FIX:
   41065: Disconnect existing CIFS sessions from the same client when a virtual circuit zero session is opened. ALF-13815
   41280: Moved session cleanup config into the base authenticator, added support to passthru/base authentication. ALF-13815
   41351: Ported database filesystem changes to fix session disconnect, from V3.4.
   41352: Merge V3.4-BUG-FIX to V4.1-BUG-FIX:
   41067: Added session disconnect support to the Alfresco CIFS authenticator. ALF-13815.
   41281: Added session cleanup support to passthru authenticator, session cleanup config moved to base class. ALF-13815 
   41353: Update svn:mergeinfo
   41355: Fix for ALF-15869 - "Site Content" dashlet shows all documents from all the sites in Alfresco Share
    - mistakenly did a record-only merge of this from 4.1->4.1.1
   41363: Fix for ALF-14875 - Manage permissions shows the permission 'No privileges' for All Other Users
    - reworked the permissions dialog and permissions panels to correctly handle multiple permissions on a special permissions group such as GROUP_EVERYONE
    - now correctly gets/sets permissions for GROUP_EVERYONE
    - this also fixes ALF-12014 - in that it allows custom SiteXYZ permissions to work correctly again also (will need manual backport for 3.4.X though)
    - removed hacks related to previous attempts to fix the above issue
    - added lots of comments around relevant sections to add in future refactoring or understanding
   41371: Merged BRANCHES/DEV/V3.4-BUG-FIX to BRANCHES/DEV/V4.1-BUG-FIX (RECORD ONLY)
      41370: Fix for ALF-12014 - Share - Custom role causes incorrect display of permissions
   Manual merge of changes to 4.1.2
   41399: ALF-13438: java.lang.OutOfMemoryError appears in alfresco log when trying to create few thousands of sites consistently. 
   - The v4.0 parent assocs cache used to store every verion of every node's parent assocs, so as we added a node to 60,000 sites, we retained all previous versions of the user's parent association map
   - After creating 10,000 sites, the cache size was about 4GB, containing about 2 million parents
   - Now we use a specialized class that allows an upper limit to be set on the total number of cached parents as well as children.
   - Because the cache is keyed by node transaction ID, the cache can be non-clustered and non-transactional
   - Once the average number of parents is more than 8, the cache will drop its oldest entries
   - ParentAssocsInfo also now uses a compact TreeMap instead of a HashMap
   41401: ALF-13438: Small correction
   41406: ALF-13438: Fix failing unit tests
   - Removed unused node.parentAssocsSharedCache and corrected node.childByNameCache not to reference it!
   - Because parent assocs are cached by transaction ID, we must always invalidate them on an in-transaction version increment
   41409: Logging of unexepected errors on FTP
   41411: Fix possible FTP data session leak if client mixes PORT and PASV commands. ALF-15126
   41412: ALF-15845 : Clone for Hotfix: Word document on Windows via CIFS becomes locked (Read Only) when network drops temporarily
   41415: Fixes: ALF-15649: Removes country locale from files with it hard coded.
   41419: ALF-14599: Removed ftp.ipv6.enabled from enterprise overlay and bundles
   41426: ALF-15845 Clone for Hotfix: Word document on Windows via CIFS becomes locked (Read Only) when network drops temporarily
     Roll back changes to DiskDriver interface in favour of hacking NetworkFile.
   41440: JAPANESE: Translation update based on EN r41099
   41446: ALF-13091: Remove unecessary bean post processors from sub ssytem context and remove CXF's Jsr250BeanPostProcessor.
   41458: RUSSIAN: Further translation updates following linguistic review.
   41459: ALF-15897: Revert revision 41446, an attempted fix for ALF-13091
   41487: Fix for ALF-15910 SOLR - Add index warming and filter pointless entries from the filter cache
   Fix for ALF-15851
   Too many live instances of SolrIndexSearcher at one time resulting in OOM - Alfresco 4.1.1 - build 151
   41506: Merged DEV to V4.1-BUG-FIX
      41505: ALF-15879: PostgreSQL: upgrade 2.2.8 (577) -> 3.4.10 (703) -> 4.1.1 (159) failed.
             - Make dropping "store_id" index and "alf_node_store_id_key" constraint optional in 4.1.1 upgrade script
               because clean 3.4 has "store_id" index and doesn't have "alf_node_store_id_key" constraint,
               but 3.4 upgraded from 2.2  has "alf_node_store_id_key" constraint and doesn't have "store_id" index.
   41531: Fixed ALF-15687, so that any user (except for Admin) won’t be able to retrieve any other user’s preferences via REST API. Also, updated the preferences controllers for the Post and Delete.
   41539: ALF-15899: Inbound email does not support multiple recipient folders
   - Fix by Dmitry Vaserin
   41540: Merged V3.4-BUG-FIX to V4.1-BUG-FIX
      40794: Merged DEV to V3.4-BUG-FIX
         40793: ALF-13752 Saving Word (mac 2011) documents via CIFS into a folder with Versionable rules on Mac OS X Lion (Fix for 3.4)
            In ContentDiskDriver.renameFile() was added a check whether a node in the archive.
      40806: Fix for ALF-9787 - Hiding sites in Share with permissionsDefinitions.xml [creates a permissions error in the blog portion of the site]
      40922: Merged DEV to V3.4-BUG-FIX (reviewed by Frederik)
         40488: ALF-13357 : Empty outcome when a timer is invoked
            A check for transitionName was added to AlfrescoTimer to support custom transitions.
      40940: ALF-15696: Remove svnkit.jar - makes the build fail if the version of installed command line svn is 1.7. Using commandline binding for <svn> Ant task to be consistent with other parts of the build.
      41066: Various fixes to the database filesystem for session disconnect cleanup.
      41068: Updates to the Alfresco filesystem for session cleanup. ALF-13815
      41301: Merged PATCHES/V3.4.9 to V3.4-BUG-FIX
         40966: ALF-15846 / ALF-15709: OOM on cascading reindex
         - Avoid buffering of all the affected PATH documents in memory - used the set of 'visited' paths to delay generation to the final flush.
         41044: ALF-15847 / ALF-15748: Lucene indexer can make sub-optimal cascade reindex decisions during an LDAP sync.
         - When a user in 6 groups was removed from a massive group, the massive group was getting cascade reindexed rather than the user
         - Logic adjusted as follows:
            For nodes with 5 or less parents, we always cascade reindex the child node. For nodes with more than 5 parents, we cascade reindex the parent node if it has less children than the child has parents.
      41395: ALF-15715: Unable to edit properties whilst transformation in progress
         - Delayed all changes that would potentially lock the parent node row while the thumbnail is generating.
         - Reorganized RenditionContext to lazily instantiate its destination node.
         - Changed render destination to use temporary ContentWriter until render is complete.
      41396: ALF-15715: Fix for failing unit tests.
      41413: Fix for issue where user calendar remote api was generating invalid date searches for user dashlet calendar.
      41509: Merged DEV to V3.4-BUG-FIX
         41507: ALF-12833: Issues installing Alfresco on WebSphere when the server doesn't have internet access
            Context-param which Sets "http://apache.org/xml/features/nonvalidating/load-external-dtd" feature on the SAXParser to false if this parameter is false 
      41510: ALF-15171: After addition of a secondary parent association to a container, not all index paths were being regenerated due to a logic error
      41512: ALF-15919: Merged PATCHES/V3.4.10 to V3.4-BUG-FIX
         41091: ALF-15723:  Merged DEV to PATCHES/V3.4.10
            26579: Switch the transformer to use Tika


git-svn-id: https://svn.alfresco.com/repos/alfresco-enterprise/alfresco/HEAD/root@41543 c4b6b30b-aa2e-2d43-bbcb-ca4b014f7261
2012-09-12 19:10:54 +00:00

1126 lines
46 KiB
Java

/*
* Copyright (C) 2005-2012 Alfresco Software Limited.
*
* This file is part of Alfresco
*
* Alfresco is free software: you can redistribute it and/or modify
* it under the terms of the GNU Lesser General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* Alfresco is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public License
* along with Alfresco. If not, see <http://www.gnu.org/licenses/>.
*/
package org.alfresco.repo.content;
import java.io.File;
import java.io.Serializable;
import java.util.ArrayList;
import java.util.Collection;
import java.util.HashMap;
import java.util.HashSet;
import java.util.List;
import java.util.Map;
import java.util.Set;
import org.alfresco.error.AlfrescoRuntimeException;
import org.alfresco.model.ContentModel;
import org.alfresco.repo.avm.AVMNodeConverter;
import org.alfresco.repo.content.ContentServicePolicies.OnContentPropertyUpdatePolicy;
import org.alfresco.repo.content.ContentServicePolicies.OnContentReadPolicy;
import org.alfresco.repo.content.ContentServicePolicies.OnContentUpdatePolicy;
import org.alfresco.repo.content.cleanup.EagerContentStoreCleaner;
import org.alfresco.repo.content.filestore.FileContentStore;
import org.alfresco.repo.content.filestore.FileContentWriter;
import org.alfresco.repo.content.transform.ContentTransformer;
import org.alfresco.repo.content.transform.ContentTransformerRegistry;
import org.alfresco.repo.content.transform.OOXMLThumbnailContentTransformer;
import org.alfresco.repo.content.transform.TransformerDebug;
import org.alfresco.repo.content.transform.UnimportantTransformException;
import org.alfresco.repo.node.NodeServicePolicies;
import org.alfresco.repo.policy.ClassPolicyDelegate;
import org.alfresco.repo.policy.JavaBehaviour;
import org.alfresco.repo.policy.PolicyComponent;
import org.alfresco.repo.transaction.RetryingTransactionHelper;
import org.alfresco.service.cmr.avm.AVMService;
import org.alfresco.service.cmr.dictionary.DataTypeDefinition;
import org.alfresco.service.cmr.dictionary.DictionaryService;
import org.alfresco.service.cmr.dictionary.InvalidTypeException;
import org.alfresco.service.cmr.dictionary.PropertyDefinition;
import org.alfresco.service.cmr.repository.ContentData;
import org.alfresco.service.cmr.repository.ContentIOException;
import org.alfresco.service.cmr.repository.ContentReader;
import org.alfresco.service.cmr.repository.ContentService;
import org.alfresco.service.cmr.repository.ContentWriter;
import org.alfresco.service.cmr.repository.MimetypeService;
import org.alfresco.service.cmr.repository.NoTransformerException;
import org.alfresco.service.cmr.repository.NodeRef;
import org.alfresco.service.cmr.repository.NodeService;
import org.alfresco.service.cmr.repository.StoreRef;
import org.alfresco.service.cmr.repository.TransformationOptions;
import org.alfresco.service.cmr.usage.ContentQuotaException;
import org.alfresco.service.namespace.QName;
import org.alfresco.util.EqualsHelper;
import org.alfresco.util.Pair;
import org.alfresco.util.TempFileProvider;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.springframework.beans.BeansException;
import org.springframework.context.ApplicationContext;
import org.springframework.context.ApplicationContextAware;
import org.springframework.extensions.surf.util.I18NUtil;
/**
* Service implementation acting as a level of indirection between the client
* and the underlying content store.
* <p>
* Note: This class was formerly the {@link RoutingContentService} but the
* 'routing' functionality has been pushed into the {@link AbstractRoutingContentStore store}
* implementations.
*
* @author Derek Hulley
* @since 3.2
*/
public class ContentServiceImpl implements ContentService, ApplicationContextAware
{
private static Log logger = LogFactory.getLog(ContentServiceImpl.class);
private DictionaryService dictionaryService;
private NodeService nodeService;
private AVMService avmService;
private MimetypeService mimetypeService;
private RetryingTransactionHelper transactionHelper;
private ApplicationContext applicationContext;
protected TransformerDebug transformerDebug;
/** a registry of all available content transformers */
private ContentTransformerRegistry transformerRegistry;
/** The cleaner that will ensure that rollbacks clean up after themselves */
private EagerContentStoreCleaner eagerContentStoreCleaner;
/** the store to use. Any multi-store support is provided by the store implementation. */
private ContentStore store;
/** the store for all temporarily created content */
private ContentStore tempStore;
private ContentTransformer imageMagickContentTransformer;
/** Should we consider zero byte content to be the same as no content? */
private boolean ignoreEmptyContent;
private boolean transformerFailover;
/**
* The policy component
*/
private PolicyComponent policyComponent;
/*
* Policies delegates
*/
ClassPolicyDelegate<ContentServicePolicies.OnContentUpdatePolicy> onContentUpdateDelegate;
ClassPolicyDelegate<ContentServicePolicies.OnContentPropertyUpdatePolicy> onContentPropertyUpdateDelegate;
ClassPolicyDelegate<ContentServicePolicies.OnContentReadPolicy> onContentReadDelegate;
public void setRetryingTransactionHelper(RetryingTransactionHelper helper)
{
this.transactionHelper = helper;
}
public void setDictionaryService(DictionaryService dictionaryService)
{
this.dictionaryService = dictionaryService;
}
public void setNodeService(NodeService nodeService)
{
this.nodeService = nodeService;
}
public void setMimetypeService(MimetypeService mimetypeService)
{
this.mimetypeService = mimetypeService;
}
public void setTransformerRegistry(ContentTransformerRegistry transformerRegistry)
{
this.transformerRegistry = transformerRegistry;
}
public void setEagerContentStoreCleaner(EagerContentStoreCleaner eagerContentStoreCleaner)
{
this.eagerContentStoreCleaner = eagerContentStoreCleaner;
}
public void setStore(ContentStore store)
{
this.store = store;
}
public void setPolicyComponent(PolicyComponent policyComponent)
{
this.policyComponent = policyComponent;
}
public void setAvmService(AVMService service)
{
this.avmService = service;
}
public void setImageMagickContentTransformer(ContentTransformer imageMagickContentTransformer)
{
this.imageMagickContentTransformer = imageMagickContentTransformer;
}
public void setIgnoreEmptyContent(boolean ignoreEmptyContent)
{
this.ignoreEmptyContent = ignoreEmptyContent;
}
/**
* Allows fail over form one transformer to another when there is
* more than one transformer available. The cost is that the output
* of the transformer must go to a temporary file in case it fails.
* @param transformerFailover {@code true} indicate that fail over
* should take place.
*/
public void setTransformerFailover(boolean transformerFailover)
{
this.transformerFailover = transformerFailover;
}
/* (non-Javadoc)
* @see org.springframework.context.ApplicationContextAware#setApplicationContext(org.springframework.context.ApplicationContext)
*/
public void setApplicationContext(ApplicationContext applicationContext) throws BeansException
{
this.applicationContext = applicationContext;
}
/**
* Helper setter of the transformer debug.
* @param transformerDebug
*/
public void setTransformerDebug(TransformerDebug transformerDebug)
{
this.transformerDebug = transformerDebug;
}
/**
* Service initialise
*/
public void init()
{
// Set up a temporary store
this.tempStore = new FileContentStore(this.applicationContext, TempFileProvider.getTempDir().getAbsolutePath());
// Bind on update properties behaviour
this.policyComponent.bindClassBehaviour(
NodeServicePolicies.OnUpdatePropertiesPolicy.QNAME,
this,
new JavaBehaviour(this, "onUpdateProperties"));
// Register on content update policy
this.onContentUpdateDelegate = this.policyComponent.registerClassPolicy(OnContentUpdatePolicy.class);
this.onContentPropertyUpdateDelegate = this.policyComponent.registerClassPolicy(OnContentPropertyUpdatePolicy.class);
this.onContentReadDelegate = this.policyComponent.registerClassPolicy(OnContentReadPolicy.class);
}
/**
* Update properties policy behaviour
*
* @param nodeRef the node reference
* @param before the before values of the properties
* @param after the after values of the properties
*/
public void onUpdateProperties(
NodeRef nodeRef,
Map<QName, Serializable> before,
Map<QName, Serializable> after)
{
// ALF-254: empty files (0 bytes) do not trigger content rules
if (nodeService.hasAspect(nodeRef, ContentModel.ASPECT_NO_CONTENT))
{
return;
}
// Don't duplicate work when firing multiple policies
Set<QName> types = null;
OnContentPropertyUpdatePolicy propertyPolicy = null; // Doesn't change for the node instance
// Variables to control firing of node-level policies (any content change)
boolean fire = false;
boolean isNewContent = false;
// check if any of the content properties have changed
for (QName propertyQName : after.keySet())
{
// is this a content property?
PropertyDefinition propertyDef = dictionaryService.getProperty(propertyQName);
if (propertyDef == null)
{
// the property is not recognised
continue;
}
else if (!propertyDef.getDataType().getName().equals(DataTypeDefinition.CONTENT))
{
// not a content type
continue;
}
else if (propertyDef.isMultiValued())
{
// We don't fire notifications for multi-valued content properties
continue;
}
try
{
ContentData beforeValue = (ContentData) before.get(propertyQName);
ContentData afterValue = (ContentData) after.get(propertyQName);
boolean hasContentBefore = ContentData.hasContent(beforeValue)
&& (!ignoreEmptyContent || beforeValue.getSize() > 0);
boolean hasContentAfter = ContentData.hasContent(afterValue)
&& (!ignoreEmptyContent || afterValue.getSize() > 0);
// There are some shortcuts here
if (!hasContentBefore && !hasContentAfter)
{
// Really, nothing happened
continue;
}
else if (EqualsHelper.nullSafeEquals(beforeValue, afterValue))
{
// Still, nothing happening
continue;
}
// Check for new content
isNewContent = !hasContentBefore && hasContentAfter;
// So debug ...
if (logger.isDebugEnabled())
{
String name = (String) nodeService.getProperty(nodeRef, ContentModel.PROP_NAME);
logger.debug(
"Content property updated: \n" +
" Node Name: " + name + "\n" +
" Property: " + propertyQName + "\n" +
" Is new: " + isNewContent + "\n" +
" Before: " + beforeValue + "\n" +
" After: " + afterValue);
}
// Fire specific policy
types = getTypes(nodeRef, types);
if (propertyPolicy == null)
{
propertyPolicy = onContentPropertyUpdateDelegate.get(nodeRef, types);
}
propertyPolicy.onContentPropertyUpdate(nodeRef, propertyQName, beforeValue, afterValue);
// We also fire an event if *any* content property is changed
fire = true;
}
catch (ClassCastException e)
{
// properties don't conform to model
continue;
}
}
// fire?
if (fire)
{
// Fire the content update policy
types = getTypes(nodeRef, types);
OnContentUpdatePolicy policy = onContentUpdateDelegate.get(nodeRef, types);
policy.onContentUpdate(nodeRef, isNewContent);
}
}
/**
* Helper method to lazily populate the types associated with a node
*
* @param nodeRef the node
* @param types any existing types
* @return the types - either newly populated or just what was passed in
*/
private Set<QName> getTypes(NodeRef nodeRef, Set<QName> types)
{
if (types != null)
{
return types;
}
types = new HashSet<QName>(this.nodeService.getAspects(nodeRef));
types.add(this.nodeService.getType(nodeRef));
return types;
}
@Override
public long getStoreFreeSpace()
{
return store.getSpaceFree();
}
@Override
public long getStoreTotalSpace()
{
return store.getSpaceTotal();
}
/** {@inheritDoc} */
public ContentReader getRawReader(String contentUrl)
{
ContentReader reader = null;
try
{
reader = store.getReader(contentUrl);
}
catch (UnsupportedContentUrlException e)
{
// The URL is not supported, so we spoof it
reader = new EmptyContentReader(contentUrl);
}
if (reader == null)
{
throw new AlfrescoRuntimeException("ContentStore implementations may not return null ContentReaders");
}
// set extra data on the reader
reader.setMimetype(MimetypeMap.MIMETYPE_BINARY);
reader.setEncoding("UTF-8");
reader.setLocale(I18NUtil.getLocale());
// Done
if (logger.isDebugEnabled())
{
logger.debug(
"Direct request for reader: \n" +
" Content URL: " + contentUrl + "\n" +
" Reader: " + reader);
}
return reader;
}
public ContentReader getReader(NodeRef nodeRef, QName propertyQName)
{
return getReader(nodeRef, propertyQName, true);
}
@SuppressWarnings("unchecked")
private ContentReader getReader(NodeRef nodeRef, QName propertyQName, boolean fireContentReadPolicy)
{
ContentData contentData = null;
Serializable propValue = nodeService.getProperty(nodeRef, propertyQName);
if (propValue instanceof Collection)
{
Collection<Serializable> colPropValue = (Collection<Serializable>)propValue;
if (colPropValue.size() > 0)
{
propValue = colPropValue.iterator().next();
}
}
if (propValue instanceof ContentData)
{
contentData = (ContentData)propValue;
}
if (contentData == null)
{
PropertyDefinition contentPropDef = dictionaryService.getProperty(propertyQName);
// if no value or a value other content, and a property definition has been provided, ensure that it's CONTENT or ANY
if (contentPropDef != null &&
(!(contentPropDef.getDataType().getName().equals(DataTypeDefinition.CONTENT) ||
contentPropDef.getDataType().getName().equals(DataTypeDefinition.ANY))))
{
throw new InvalidTypeException("The node property must be of type content: \n" +
" node: " + nodeRef + "\n" +
" property name: " + propertyQName + "\n" +
" property type: " + ((contentPropDef == null) ? "unknown" : contentPropDef.getDataType()),
propertyQName);
}
}
// check that the URL is available
if (contentData == null || contentData.getContentUrl() == null)
{
// there is no URL - the interface specifies that this is not an error condition
return null;
}
String contentUrl = contentData.getContentUrl();
// The context of the read is entirely described by the URL
ContentReader reader = store.getReader(contentUrl);
if (reader == null)
{
throw new AlfrescoRuntimeException("ContentStore implementations may not return null ContentReaders");
}
// set extra data on the reader
reader.setMimetype(contentData.getMimetype());
reader.setEncoding(contentData.getEncoding());
reader.setLocale(contentData.getLocale());
// Fire the content read policy
if (reader != null && fireContentReadPolicy == true)
{
// Fire the content update policy
Set<QName> types = new HashSet<QName>(this.nodeService.getAspects(nodeRef));
types.add(this.nodeService.getType(nodeRef));
OnContentReadPolicy policy = this.onContentReadDelegate.get(nodeRef, types);
policy.onContentRead(nodeRef);
}
// we don't listen for anything
// result may be null - but interface contract says we may return null
return reader;
}
public ContentWriter getWriter(NodeRef nodeRef, QName propertyQName, boolean update)
{
if (nodeRef == null)
{
ContentContext ctx = new ContentContext(null, null);
// for this case, we just give back a valid URL into the content store
ContentWriter writer = store.getWriter(ctx);
// Register the new URL for rollback cleanup
eagerContentStoreCleaner.registerNewContentUrl(writer.getContentUrl());
// done
return writer;
}
// check for an existing URL - the get of the reader will perform type checking
ContentReader existingContentReader = getReader(nodeRef, propertyQName, false);
// get the content using the (potentially) existing content - the new content
// can be wherever the store decides.
ContentContext ctx = new NodeContentContext(existingContentReader, null, nodeRef, propertyQName);
ContentWriter writer = store.getWriter(ctx);
// Register the new URL for rollback cleanup
eagerContentStoreCleaner.registerNewContentUrl(writer.getContentUrl());
// Special case for AVM repository.
Serializable contentValue = null;
if (nodeRef.getStoreRef().getProtocol().equals(StoreRef.PROTOCOL_AVM))
{
Pair<Integer, String> avmVersionPath = AVMNodeConverter.ToAVMVersionPath(nodeRef);
contentValue = avmService.getContentDataForWrite(avmVersionPath.getSecond());
}
else
{
contentValue = nodeService.getProperty(nodeRef, propertyQName);
}
// set extra data on the reader if the property is pre-existing
if (contentValue != null && contentValue instanceof ContentData)
{
ContentData contentData = (ContentData)contentValue;
writer.setMimetype(contentData.getMimetype());
writer.setEncoding(contentData.getEncoding());
writer.setLocale(contentData.getLocale());
}
// attach a listener if required
if (update)
{
// need a listener to update the node when the stream closes
WriteStreamListener listener = new WriteStreamListener(nodeService, nodeRef, propertyQName, writer);
listener.setRetryingTransactionHelper(transactionHelper);
writer.addListener(listener);
}
// supply the writer with a copy of the mimetype service if needed
if (writer instanceof AbstractContentWriter)
{
((AbstractContentWriter)writer).setMimetypeService(mimetypeService);
}
// give back to the client
return writer;
}
/**
* @return Returns a writer to an anonymous location
*/
public ContentWriter getTempWriter()
{
// there is no existing content and we don't specify the location of the new content
return tempStore.getWriter(ContentContext.NULL_CONTEXT);
}
/**
* @see org.alfresco.repo.content.transform.ContentTransformerRegistry
* @see org.alfresco.repo.content.transform.ContentTransformer
* @see org.alfresco.service.cmr.repository.ContentService#transform(org.alfresco.service.cmr.repository.ContentReader, org.alfresco.service.cmr.repository.ContentWriter)
*/
public void transform(ContentReader reader, ContentWriter writer)
{
// Call transform with no options
TransformationOptions options = new TransformationOptions();
this.transform(reader, writer, options);
}
/**
* @see org.alfresco.repo.content.transform.ContentTransformerRegistry
* @see org.alfresco.repo.content.transform.ContentTransformer
* @deprecated
*/
public void transform(ContentReader reader, ContentWriter writer, Map<String, Object> options)
throws NoTransformerException, ContentIOException
{
transform(reader, writer, new TransformationOptions(options));
}
/**
* @see org.alfresco.repo.content.transform.ContentTransformerRegistry
* @see org.alfresco.repo.content.transform.ContentTransformer
*/
public void transform(ContentReader reader, ContentWriter writer, TransformationOptions options)
throws NoTransformerException, ContentIOException
{
// check that source and target mimetypes are available
if (reader == null)
{
throw new AlfrescoRuntimeException("The content reader must be set");
}
String sourceMimetype = reader.getMimetype();
if (sourceMimetype == null)
{
throw new AlfrescoRuntimeException("The content reader mimetype must be set: " + reader);
}
String targetMimetype = writer.getMimetype();
if (targetMimetype == null)
{
throw new AlfrescoRuntimeException("The content writer mimetype must be set: " + writer);
}
long sourceSize = reader.getSize();
try
{
// look for a transformer
transformerDebug.pushAvailable(reader.getContentUrl(), sourceMimetype, targetMimetype, options);
List<ContentTransformer> transformers = getActiveTransformers(sourceMimetype, sourceSize, targetMimetype, options);
transformerDebug.availableTransformers(transformers, sourceSize, "ContentService.transform(...)");
int count = transformers.size();
if (count == 0)
{
throw new NoTransformerException(sourceMimetype, targetMimetype);
}
if (count == 1 || !transformerFailover)
{
ContentTransformer transformer = transformers.size() == 0 ? null : transformers.get(0);
transformer.transform(reader, writer, options);
}
else
{
failoverTransformers(reader, writer, options, targetMimetype, transformers);
}
}
finally
{
if (transformerDebug.isEnabled())
{
transformerDebug.popAvailable();
debugActiveTransformers(sourceMimetype, targetMimetype, sourceSize, options);
}
}
}
private void failoverTransformers(ContentReader reader, ContentWriter writer,
TransformationOptions options, String targetMimetype,
List<ContentTransformer> transformers)
{
List<AlfrescoRuntimeException> exceptions = null;
boolean done = false;
try
{
// Try the best transformer and then the next if it fails
// and so on down the list
char c = 'a';
String outputFileExt = mimetypeService.getExtension(targetMimetype);
for (ContentTransformer transformer : transformers)
{
ContentWriter currentWriter = writer;
File tempFile = null;
try
{
// We can't know in advance which of the
// available transformer will work - if any.
// We can't write into the ContentWriter stream.
// So make a temporary file writer with the
// current transformer name.
tempFile = TempFileProvider.createTempFile(
"FailoverTransformer_intermediate_"
+ transformer.getClass().getSimpleName() + "_", "."
+ outputFileExt);
currentWriter = new FileContentWriter(tempFile);
currentWriter.setMimetype(targetMimetype);
currentWriter.setEncoding(writer.getEncoding());
if (c != 'a' && transformerDebug.isEnabled())
{
transformerDebug.debug("");
transformerDebug.debug("Try " + c + ")");
}
c++;
transformer.transform(reader, currentWriter, options);
if (tempFile != null)
{
writer.putContent(tempFile);
}
// No need to close input or output streams
// (according
// to comment in FailoverContentTransformer)
done = true;
return;
}
catch (AlfrescoRuntimeException e)
{
if (exceptions == null)
{
exceptions = new ArrayList<AlfrescoRuntimeException>();
}
exceptions.add(e);
// Set a new reader to refresh the input stream.
reader = reader.getReader();
}
}
// Throw the exception from the first transformer. The
// others are consumed.
if (exceptions != null)
{
throw exceptions.get(0);
}
}
finally
{
// Log exceptions that we have consumed. We may have thrown the first one if
// none of the transformers worked.
if (exceptions != null)
{
boolean first = true;
for (Exception e : exceptions)
{
Throwable rootCause = (e instanceof AlfrescoRuntimeException) ? ((AlfrescoRuntimeException)e).getRootCause() : null;
String message = (rootCause == null ? null : rootCause.getMessage());
if (done)
{
message = "Transformer succeeded after previous transformer failed"+ (message == null ? "" : ": "+message);
if (rootCause instanceof UnimportantTransformException)
{
logger.debug(message);
}
else
{
logger.error(message, e);
}
}
else if (!first) // The first exception is logged later
{
message = "Transformer exception"+ (message == null ? "" : ": "+message);
if (rootCause instanceof UnimportantTransformException)
{
logger.debug(message);
}
else
{
logger.error(message, e);
}
first = false;
}
}
}
}
}
/**
* @see org.alfresco.repo.content.transform.ContentTransformerRegistry
* @see org.alfresco.repo.content.transform.ContentTransformer
*/
public ContentTransformer getTransformer(String sourceMimetype, String targetMimetype)
{
return getTransformer(null, sourceMimetype, -1, targetMimetype, new TransformationOptions());
}
public ContentTransformer getTransformer(String sourceMimetype, String targetMimetype, TransformationOptions options)
{
return getTransformer(null, sourceMimetype, -1, targetMimetype, options);
}
/**
* @see org.alfresco.service.cmr.repository.ContentService#getTransformer(String, java.lang.String, long, java.lang.String, org.alfresco.service.cmr.repository.TransformationOptions)
*/
public ContentTransformer getTransformer(String sourceUrl, String sourceMimetype, long sourceSize, String targetMimetype, TransformationOptions options)
{
List<ContentTransformer> transformers = getTransformers(sourceUrl, sourceMimetype, sourceSize, targetMimetype, options);
return (transformers == null) ? null : transformers.get(0);
}
/**
* @see org.alfresco.service.cmr.repository.ContentService#getTransformers(String, java.lang.String, long, java.lang.String, org.alfresco.service.cmr.repository.TransformationOptions)
*/
public List<ContentTransformer> getTransformers(String sourceUrl, String sourceMimetype, long sourceSize, String targetMimetype, TransformationOptions options)
{
try
{
// look for a transformer
transformerDebug.pushAvailable(sourceUrl, sourceMimetype, targetMimetype, options);
List<ContentTransformer> transformers = getActiveTransformers(sourceMimetype, sourceSize, targetMimetype, options);
transformerDebug.availableTransformers(transformers, sourceSize, "ContentService.getTransformer(...)");
return transformers.isEmpty() ? null : transformers;
}
finally
{
transformerDebug.popAvailable();
}
}
/**
* Checks if the file just uploaded into Share is a special "debugTransformers.txt" file and
* if it is creates TransformerDebug that lists all the supported mimetype transformation for
* each transformer.
*/
private void debugActiveTransformers(String sourceMimetype, String targetMimetype,
long sourceSize, TransformationOptions transformOptions)
{
// check the file name, but do faster tests first
if (sourceSize == 18 &&
MimetypeMap.MIMETYPE_TEXT_PLAIN.equals(sourceMimetype) &&
MimetypeMap.MIMETYPE_IMAGE_PNG.equals(targetMimetype) &&
"debugTransformers.txt".equals(transformerDebug.getFileName(transformOptions, true, 0)))
{
Map<String, Set<String>> explicitTransforms = debugExplicitTransforms();
debugActiveTransformersByTransformer(explicitTransforms);
debugActiveTransformersByMimetypes(explicitTransforms);
}
}
/**
* Creates TransformerDebug that lists all the supported mimetype transformation for each transformer.
*/
private void debugActiveTransformersByTransformer(Map<String, Set<String>> explicitTransforms)
{
try
{
transformerDebug.pushMisc();
transformerDebug.debug("Active and inactive transformers");
TransformationOptions options = new TransformationOptions();
for (ContentTransformer transformer: transformerRegistry.getTransformers())
{
try
{
transformerDebug.pushMisc();
int mimetypePairCount = 0;
boolean first = true;
for (String sourceMimetype : mimetypeService.getMimetypes())
{
for (String targetMimetype : mimetypeService.getMimetypes())
{
if (transformer.isTransformable(sourceMimetype, -1, targetMimetype, options))
{
long maxSourceSizeKBytes = transformer.getMaxSourceSizeKBytes(
sourceMimetype, targetMimetype, options);
// Is this an explicit transform, ignored because there are explicit transforms
// or does not have explicit transforms.
Boolean explicit = transformer.isExplicitTransformation(sourceMimetype,
targetMimetype, options);
if (!explicit)
{
Set<String> targetMimetypes = explicitTransforms.get(sourceMimetype);
explicit = (targetMimetypes == null || !targetMimetypes.contains(targetMimetype))
? null
: Boolean.FALSE;
}
transformerDebug.activeTransformer(++mimetypePairCount, transformer,
sourceMimetype, targetMimetype, maxSourceSizeKBytes, explicit, first);
first = false;
}
}
}
if (first)
{
transformerDebug.inactiveTransformer(transformer);
}
}
finally
{
transformerDebug.popMisc();
}
}
}
finally
{
transformerDebug.popMisc();
}
}
/**
* Creates TransformerDebug that lists all available transformers for each mimetype combination.
*/
private void debugActiveTransformersByMimetypes(Map<String, Set<String>> explicitTransforms)
{
try
{
transformerDebug.pushMisc();
transformerDebug.debug("Transformers for each mimetype combination");
TransformationOptions options = new TransformationOptions();
for (String sourceMimetype : mimetypeService.getMimetypes())
{
for (String targetMimetype : mimetypeService.getMimetypes())
{
try
{
transformerDebug.pushMisc();
int transformerCount = 0;
for (ContentTransformer transformer: transformerRegistry.getTransformers())
{
if (transformer.isTransformable(sourceMimetype, -1, targetMimetype, options))
{
long maxSourceSizeKBytes = transformer.getMaxSourceSizeKBytes(
sourceMimetype, targetMimetype, options);
// Is this an explicit transform, ignored because there are explicit transforms
// or does not have explicit transforms.
Boolean explicit = transformer.isExplicitTransformation(sourceMimetype,
targetMimetype, options);
if (!explicit)
{
Set<String> targetMimetypes = explicitTransforms.get(sourceMimetype);
explicit = (targetMimetypes == null || !targetMimetypes.contains(targetMimetype))
? null
: Boolean.FALSE;
}
transformerDebug.activeTransformer(sourceMimetype, targetMimetype,
transformerCount, transformer, maxSourceSizeKBytes, explicit,
transformerCount++ == 0);
}
}
}
finally
{
transformerDebug.popMisc();
}
}
}
}
finally
{
transformerDebug.popMisc();
}
}
/**
* Returns the explicit mimetype transformations. Key is the source mimetype
* and the value is a set of target mimetypes that are explicit.
*/
private Map<String, Set<String>> debugExplicitTransforms()
{
Map<String, Set<String>> explicitTransforms = new HashMap<String, Set<String>>();
TransformationOptions options = new TransformationOptions();
for (String sourceMimetype : mimetypeService.getMimetypes())
{
for (String targetMimetype : mimetypeService.getMimetypes())
{
for (ContentTransformer transformer : transformerRegistry.getTransformers())
{
if (transformer.isTransformable(sourceMimetype, -1, targetMimetype, options))
{
if (transformer.isExplicitTransformation(sourceMimetype, targetMimetype,
options))
{
Set<String> targetMimetypes = explicitTransforms.get(sourceMimetype);
if (targetMimetypes == null)
{
targetMimetypes = new HashSet<String>();
explicitTransforms.put(sourceMimetype, targetMimetypes);
}
targetMimetypes.add(targetMimetype);
break;
}
}
}
}
}
return explicitTransforms;
}
/**
* {@inheritDoc}
*/
public long getMaxSourceSizeBytes(String sourceMimetype, String targetMimetype, TransformationOptions options)
{
try
{
long maxSourceSize = 0;
transformerDebug.pushAvailable(null, sourceMimetype, targetMimetype, options);
List<ContentTransformer> transformers = getActiveTransformers(sourceMimetype, -1, targetMimetype, options);
for (ContentTransformer transformer: transformers)
{
long maxSourceSizeKBytes = transformer.getMaxSourceSizeKBytes(sourceMimetype, targetMimetype, options);
if (maxSourceSize >= 0)
{
if (maxSourceSizeKBytes < 0)
{
maxSourceSize = -1;
}
else if (maxSourceSizeKBytes > 0 && maxSourceSize < maxSourceSizeKBytes)
{
maxSourceSize = maxSourceSizeKBytes;
}
}
// if maxSourceSizeKBytes == 0 this implies the transformation is disabled
}
if (transformerDebug.isEnabled())
{
transformerDebug.availableTransformers(transformers, -1,
"ContentService.getMaxSourceSizeBytes() = "+transformerDebug.fileSize(maxSourceSize*1024));
}
return (maxSourceSize > 0) ? maxSourceSize * 1024 : maxSourceSize;
}
finally
{
transformerDebug.popAvailable();
}
}
public List<ContentTransformer> getActiveTransformers(String sourceMimetype, String targetMimetype, TransformationOptions options)
{
return getActiveTransformers(sourceMimetype, -1, targetMimetype, options);
}
public List<ContentTransformer> getActiveTransformers(String sourceMimetype, long sourceSize, String targetMimetype, TransformationOptions options)
{
return transformerRegistry.getActiveTransformers(sourceMimetype, sourceSize, targetMimetype, options);
}
/**
* @see org.alfresco.service.cmr.repository.ContentService#getImageTransformer()
*/
public ContentTransformer getImageTransformer()
{
return imageMagickContentTransformer;
}
/**
* @see org.alfresco.repo.content.transform.ContentTransformerRegistry
* @see org.alfresco.repo.content.transform.ContentTransformer
*/
public boolean isTransformable(ContentReader reader, ContentWriter writer)
{
return isTransformable(reader, writer, new TransformationOptions());
}
/**
* @see org.alfresco.service.cmr.repository.ContentService#isTransformable(org.alfresco.service.cmr.repository.ContentReader, org.alfresco.service.cmr.repository.ContentWriter, org.alfresco.service.cmr.repository.TransformationOptions)
*/
public boolean isTransformable(ContentReader reader, ContentWriter writer, TransformationOptions options)
{
// check that source and target mimetypes are available
String sourceMimetype = reader.getMimetype();
if (sourceMimetype == null)
{
throw new AlfrescoRuntimeException("The content reader mimetype must be set: " + reader);
}
String targetMimetype = writer.getMimetype();
if (targetMimetype == null)
{
throw new AlfrescoRuntimeException("The content writer mimetype must be set: " + writer);
}
long sourceSize = reader.getSize();
try
{
// look for a transformer
transformerDebug.pushAvailable(reader.getContentUrl(), sourceMimetype, targetMimetype, options);
List<ContentTransformer> transformers = getActiveTransformers(sourceMimetype, sourceSize, targetMimetype, options);
transformerDebug.availableTransformers(transformers, sourceSize, "ContentService.isTransformable(...)");
return transformers.size() > 0;
}
finally
{
transformerDebug.popAvailable();
}
}
/**
* Ensures that, upon closure of the output stream, the node is updated with
* the latest URL of the content to which it refers.
* <p>
*
* @author Derek Hulley
*/
private static class WriteStreamListener extends AbstractContentStreamListener
{
private NodeService nodeService;
private NodeRef nodeRef;
private QName propertyQName;
private ContentWriter writer;
public WriteStreamListener(
NodeService nodeService,
NodeRef nodeRef,
QName propertyQName,
ContentWriter writer)
{
this.nodeService = nodeService;
this.nodeRef = nodeRef;
this.propertyQName = propertyQName;
this.writer = writer;
}
public void contentStreamClosedImpl() throws ContentIOException
{
try
{
// set the full content property
ContentData contentData = writer.getContentData();
// Bypass NodeService for avm stores.
if (nodeRef.getStoreRef().getProtocol().equals(StoreRef.PROTOCOL_AVM))
{
nodeService.setProperty(nodeRef, ContentModel.PROP_CONTENT, contentData);
}
else
{
nodeService.setProperty(
nodeRef,
propertyQName,
contentData);
}
// done
if (logger.isDebugEnabled())
{
logger.debug("Stream listener updated node: \n" +
" node: " + nodeRef + "\n" +
" property: " + propertyQName + "\n" +
" value: " + contentData);
}
}
catch (ContentQuotaException qe)
{
throw qe;
}
catch (Throwable e)
{
throw new ContentIOException("Failed to set content property on stream closure: \n" +
" node: " + nodeRef + "\n" +
" property: " + propertyQName + "\n" +
" writer: " + writer + "\n" +
e.toString(),
e);
}
}
}
}