mirror of
https://github.com/Alfresco/alfresco-community-repo.git
synced 2025-06-30 18:15:39 +00:00
20954: Calendar Dashlet updates. Fixes: ALF-2907 (meeting workspace issue) 20968: Minor VersionableAspect fix to onDeleteNode policy firing (follow-on for r19507) 20981: Removed Install Jammer installers from V3.3.3 20982: AVMTestSuite - run PurgeTestP after AVMServiceTest (investigating ALF-3611) 20997: Fix for ALF-2605 - updated share-config-custom.xml.sample and removed obsolete extension files 21030: Pulled XAMConnector AMP into main 3.3SP3 codeline. Apart from build changes (incl. EOL), there are no repo changes. 21032: StoreSelector passes through NodeContentContext allowing stores access to node information 21036: Fix ALF-245: Can't delete space that contains "translation without content" - Delete triggers 'unmakeTranslation' - Empty translations are marked with sys:temporary before being deleted 21051: More on fix ALF-245. Reduced complexity by not deleting empty translations twice 21064: Merged V3.3 to V3.3-BUG_FIX 20898: Merged HEAD to V3.3 20724: AVMTestSuite - temporarily comment out PurgeTestP - TODO: investigate intermittent test failure 20903: Incremented version revision 20921: AVM - fix purge store so that vr nodes are actually orphaned (ALF-3627) 20952: Fix for ALF-3704: Module conflict - Alfresco web client config property page missing metadata. This is application of a fix made to the config service in a hotfix. The change provides a deterministic load order for config files loaded via the ConfigBootstrap spring bean. More importantly it means that config files loaded by modules on different machines in a cluster load in the same order. The forms client and AWE config files have been updated to take advantage of the new loading order. 21061: Merged PATCHES/V3.1.2 to V3.3 (RECORD ONLY) 20890: ALF-3687: Apply LUCENE-1383 patch to Lucene 2.1.0 to reduce memory leaks from ThreadLocals 20891: ALF-3687: Build classpath fix 20892: Incremented version label 21062: Merged PATCHES/V3.2.1 to V3.3 20897: (RECORD ONLY) Incremented version label 20901: (RECORD ONLY) ALF-3740: Merged V3.3 to PATCHES/V3.2.1 20524: VersionMigrator - option to run as scheduled job (ALF-1000) 20904: (RECORD ONLY) ALF-3732: Merged PATCHES/V3.2.r to PATCHES/V3.2.1 19803: ALF-558: File servers (CIFS / FTP / NFS) can now handle concurrent write operations on Alfresco repository - ContentDiskDriver / AVMDiskDriver now use retrying transactions for write operations - Disable EagerContentStoreCleaner on ContentDiskDriver / AVMDiskDriver closeFile() operations so that they may be retried after rollback (Sony zero byte problem) - Allow manual association of AVM ContentData with nodes so that closeFile() may be retried - Propagation of new argument through AVM interfaces 20905: (RECORD ONLY) ALF-3732: Rolled back the now unnecessary reference()/dereference() stuff from ALF-558 20906: (RECORD ONLY) ALF-3732: Merged DEV/V3.3-BUG-FIX to PATCHES/V3.2.1 20623: Fix for ALF-3188 : Access Denied when updating doc via CIFS 20907: (RECORD ONLY) ALF-3732: Merged V3.3 to PATCHES/V3.2.1 20173: Propagate IOExceptions from retryable write transactions in AlfrescoDiskDriver 20950: ALF-3779: Upgrades on large repositories from v2.1 and v2.2 were failing on MySQL due to "The total number of locks exceeds the lock table size" errors - Solution was to add support for new --BEGIN TXN and --END TXN comments and execute LOCK TABLES statements in the same transaction as large INSERT - SELECT statements. 20990: ALF-3789: Concurrency issues with InMemoryTicketComponentImpl - Previous ETHREEOH-1842 method of caching web session 'ref counts' against tickets could cause tickets to unpredictably fall out of the transactional cache - Rolled back original ETHREEOH-1842 fix. Would be too much overhead to keep these ref counts consistent across a cluster. - Instead, avoid invalidating tickets on web session timeout and only do it on explicit log out. - Now tickets maintained in non-transactional shared cache so they can't drop out unpredictably - Logic for ticket inactivity timeout caching improved so that it should work across a cluster 20991: (RECORD ONLY) Incremented version label 20993: ALF-3789: Fixed Spring configuration backward compatibily issue with previous fix - Ticket cache bean name restored to ticketsCache. This is actually now a non-transactional cache. - Also externalized parameters so that they can now be controlled by alfresco-global.properties without any bean overrides authentication.ticket.ticketsExpire=false authentication.ticket.expiryMode=AFTER_FIXED_TIME authentication.ticket.validDuration=PT1H 20994: Eclipse classpath fixes for unit testing after ant build 21057: ALF-3592: PassthruCifsAuthenticator now auto-creates / imports users who do not already exist in Alfresco - At least one of the following properties must be true for this to happen synchronization.autoCreatePeopleOnLogin synchronization.syncWhenMissingPeopleLogIn - Also improved debug logging of unknown passthru domains 21063: Merged PATCHES/V3.2.r to V3.3 21037: ALF-3793: Final attempt at realigning saved XForm data with a modified Schema - removeRemovedNodes / insertUpdatedNodes / insertPrototypeNodes replaced by a one stop recursive process that builds a new instance tree from scratch - Nodes copied over in correct order - Missing nodes added in and extra nodes discarded - Prototype nodes appended at appropriate points 21038: (RECORD ONLY) Incremented version label git-svn-id: https://svn.alfresco.com/repos/alfresco-enterprise/alfresco/HEAD/root@21065 c4b6b30b-aa2e-2d43-bbcb-ca4b014f7261
1673 lines
65 KiB
Java
1673 lines
65 KiB
Java
/*
|
|
* Copyright (C) 2005-2010 Alfresco Software Limited.
|
|
*
|
|
* This file is part of Alfresco
|
|
*
|
|
* Alfresco is free software: you can redistribute it and/or modify
|
|
* it under the terms of the GNU Lesser General Public License as published by
|
|
* the Free Software Foundation, either version 3 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* Alfresco is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with Alfresco. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
package org.alfresco.repo.domain.schema;
|
|
|
|
import java.io.BufferedReader;
|
|
import java.io.BufferedWriter;
|
|
import java.io.File;
|
|
import java.io.FileInputStream;
|
|
import java.io.FileWriter;
|
|
import java.io.InputStream;
|
|
import java.io.InputStreamReader;
|
|
import java.io.Writer;
|
|
import java.sql.Connection;
|
|
import java.sql.DatabaseMetaData;
|
|
import java.sql.DriverManager;
|
|
import java.sql.ResultSet;
|
|
import java.sql.SQLException;
|
|
import java.sql.Statement;
|
|
import java.sql.Types;
|
|
import java.util.ArrayList;
|
|
import java.util.HashMap;
|
|
import java.util.List;
|
|
import java.util.Map;
|
|
import java.util.Properties;
|
|
|
|
import org.alfresco.error.AlfrescoRuntimeException;
|
|
import org.alfresco.ibatis.SerializableTypeHandlerCallback;
|
|
import org.alfresco.repo.admin.patch.Patch;
|
|
import org.alfresco.repo.admin.patch.impl.SchemaUpgradeScriptPatch;
|
|
import org.alfresco.repo.content.filestore.FileContentWriter;
|
|
import org.alfresco.repo.domain.PropertyValue;
|
|
import org.alfresco.repo.domain.hibernate.dialect.AlfrescoOracle9Dialect;
|
|
import org.alfresco.repo.domain.hibernate.dialect.AlfrescoSQLServerDialect;
|
|
import org.alfresco.repo.domain.hibernate.dialect.AlfrescoSybaseAnywhereDialect;
|
|
import org.alfresco.service.ServiceRegistry;
|
|
import org.alfresco.service.cmr.repository.ContentWriter;
|
|
import org.alfresco.service.descriptor.Descriptor;
|
|
import org.alfresco.service.descriptor.DescriptorService;
|
|
import org.alfresco.util.ApplicationContextHelper;
|
|
import org.alfresco.util.LogUtil;
|
|
import org.alfresco.util.TempFileProvider;
|
|
import org.alfresco.util.schemadump.Main;
|
|
import org.apache.commons.logging.Log;
|
|
import org.apache.commons.logging.LogFactory;
|
|
import org.hibernate.HibernateException;
|
|
import org.hibernate.Session;
|
|
import org.hibernate.SessionFactory;
|
|
import org.hibernate.cfg.Configuration;
|
|
import org.hibernate.cfg.Environment;
|
|
import org.hibernate.connection.UserSuppliedConnectionProvider;
|
|
import org.hibernate.dialect.DB2Dialect;
|
|
import org.hibernate.dialect.DerbyDialect;
|
|
import org.hibernate.dialect.Dialect;
|
|
import org.hibernate.dialect.HSQLDialect;
|
|
import org.hibernate.dialect.MySQL5Dialect;
|
|
import org.hibernate.dialect.MySQLDialect;
|
|
import org.hibernate.dialect.MySQLInnoDBDialect;
|
|
import org.hibernate.dialect.Oracle10gDialect;
|
|
import org.hibernate.dialect.Oracle9Dialect;
|
|
import org.hibernate.dialect.Oracle9iDialect;
|
|
import org.hibernate.dialect.OracleDialect;
|
|
import org.hibernate.dialect.PostgreSQLDialect;
|
|
import org.hibernate.engine.ActionQueue;
|
|
import org.hibernate.tool.hbm2ddl.DatabaseMetadata;
|
|
import org.hibernate.tool.hbm2ddl.SchemaExport;
|
|
import org.springframework.context.ApplicationContext;
|
|
import org.springframework.context.ApplicationEvent;
|
|
import org.springframework.core.io.Resource;
|
|
import org.springframework.core.io.support.PathMatchingResourcePatternResolver;
|
|
import org.springframework.core.io.support.ResourcePatternResolver;
|
|
import org.springframework.extensions.surf.util.AbstractLifecycleBean;
|
|
import org.springframework.extensions.surf.util.I18NUtil;
|
|
import org.springframework.orm.hibernate3.LocalSessionFactoryBean;
|
|
|
|
/**
|
|
* Bootstraps the schema and schema update. The schema is considered missing if the applied patch table
|
|
* is not present, and the schema is considered empty if the applied patch table is empty.
|
|
*
|
|
* @author Derek Hulley
|
|
*/
|
|
public class SchemaBootstrap extends AbstractLifecycleBean
|
|
{
|
|
/** The placeholder for the configured <code>Dialect</code> class name: <b>${db.script.dialect}</b> */
|
|
private static final String PLACEHOLDER_SCRIPT_DIALECT = "\\$\\{db\\.script\\.dialect\\}";
|
|
|
|
private static final String MSG_DIALECT_USED = "schema.update.msg.dialect_used";
|
|
private static final String MSG_BYPASSING_SCHEMA_UPDATE = "schema.update.msg.bypassing";
|
|
private static final String MSG_NORMALIZED_SCHEMA = "schema.update.msg.normalized_schema";
|
|
private static final String MSG_NORMALIZED_SCHEMA_PRE = "schema.update.msg.normalized_schema_pre";
|
|
private static final String MSG_NORMALIZED_SCHEMA_POST = "schema.update.msg.normalized_schema_post";
|
|
private static final String MSG_NO_CHANGES = "schema.update.msg.no_changes";
|
|
private static final String MSG_ALL_STATEMENTS = "schema.update.msg.all_statements";
|
|
private static final String MSG_EXECUTING_GENERATED_SCRIPT = "schema.update.msg.executing_generated_script";
|
|
private static final String MSG_EXECUTING_COPIED_SCRIPT = "schema.update.msg.executing_copied_script";
|
|
private static final String MSG_EXECUTING_STATEMENT = "schema.update.msg.executing_statement";
|
|
private static final String MSG_OPTIONAL_STATEMENT_FAILED = "schema.update.msg.optional_statement_failed";
|
|
private static final String WARN_DIALECT_UNSUPPORTED = "schema.update.warn.dialect_unsupported";
|
|
private static final String WARN_DIALECT_SUBSTITUTING = "schema.update.warn.dialect_substituting";
|
|
private static final String WARN_DIALECT_HSQL = "schema.update.warn.dialect_hsql";
|
|
private static final String WARN_DIALECT_DERBY = "schema.update.warn.dialect_derby";
|
|
private static final String ERR_FORCED_STOP = "schema.update.err.forced_stop";
|
|
private static final String ERR_DIALECT_SHOULD_USE = "schema.update.err.dialect_should_use";
|
|
private static final String ERR_MULTIPLE_SCHEMAS = "schema.update.err.found_multiple";
|
|
private static final String ERR_PREVIOUS_FAILED_BOOTSTRAP = "schema.update.err.previous_failed";
|
|
private static final String ERR_STATEMENT_FAILED = "schema.update.err.statement_failed";
|
|
private static final String ERR_UPDATE_FAILED = "schema.update.err.update_failed";
|
|
private static final String ERR_VALIDATION_FAILED = "schema.update.err.validation_failed";
|
|
private static final String ERR_SCRIPT_NOT_RUN = "schema.update.err.update_script_not_run";
|
|
private static final String ERR_SCRIPT_NOT_FOUND = "schema.update.err.script_not_found";
|
|
private static final String ERR_STATEMENT_VAR_ASSIGNMENT_BEFORE_SQL = "schema.update.err.statement_var_assignment_before_sql";
|
|
private static final String ERR_STATEMENT_VAR_ASSIGNMENT_FORMAT = "schema.update.err.statement_var_assignment_format";
|
|
private static final String ERR_STATEMENT_TERMINATOR = "schema.update.err.statement_terminator";
|
|
|
|
public static final int DEFAULT_LOCK_RETRY_COUNT = 24;
|
|
public static final int DEFAULT_LOCK_RETRY_WAIT_SECONDS = 5;
|
|
|
|
public static final int DEFAULT_MAX_STRING_LENGTH = 1024;
|
|
private static volatile int maxStringLength = DEFAULT_MAX_STRING_LENGTH;
|
|
private Dialect dialect;
|
|
|
|
private ResourcePatternResolver rpr = new PathMatchingResourcePatternResolver(this.getClass().getClassLoader());
|
|
|
|
|
|
/**
|
|
* @see PropertyValue#DEFAULT_MAX_STRING_LENGTH
|
|
*/
|
|
private static final void setMaxStringLength(int length)
|
|
{
|
|
if (length < 1024)
|
|
{
|
|
throw new AlfrescoRuntimeException("The maximum string length must >= 1024 characters.");
|
|
}
|
|
SchemaBootstrap.maxStringLength = length;
|
|
}
|
|
|
|
/**
|
|
* @return Returns the maximum number of characters that a string field can be
|
|
*/
|
|
public static final int getMaxStringLength()
|
|
{
|
|
return SchemaBootstrap.maxStringLength;
|
|
}
|
|
|
|
/**
|
|
* Truncates or returns a string that will fit into the string columns in the schema. Text fields can
|
|
* either cope with arbitrarily long text fields or have the default limit, {@link #DEFAULT_MAX_STRING_LENGTH}.
|
|
*
|
|
* @param value the string to check
|
|
* @return Returns a string that is short enough for {@link SchemaBootstrap#getMaxStringLength()}
|
|
*
|
|
* @since 3.2
|
|
*/
|
|
public static final String trimStringForTextFields(String value)
|
|
{
|
|
if (value != null && value.length() > maxStringLength)
|
|
{
|
|
return value.substring(0, maxStringLength);
|
|
}
|
|
else
|
|
{
|
|
return value;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Sets the previously auto-detected Hibernate dialect.
|
|
*
|
|
* @param dialect
|
|
* the dialect
|
|
*/
|
|
public void setDialect(Dialect dialect)
|
|
{
|
|
this.dialect = dialect;
|
|
}
|
|
|
|
private static Log logger = LogFactory.getLog(SchemaBootstrap.class);
|
|
|
|
private LocalSessionFactoryBean localSessionFactory;
|
|
private String schemaOuputFilename;
|
|
private boolean updateSchema;
|
|
private boolean stopAfterSchemaBootstrap;
|
|
private List<String> preCreateScriptUrls;
|
|
private List<String> postCreateScriptUrls;
|
|
private List<SchemaUpgradeScriptPatch> validateUpdateScriptPatches;
|
|
private List<SchemaUpgradeScriptPatch> preUpdateScriptPatches;
|
|
private List<SchemaUpgradeScriptPatch> postUpdateScriptPatches;
|
|
private int schemaUpdateLockRetryCount = DEFAULT_LOCK_RETRY_COUNT;
|
|
private int schemaUpdateLockRetryWaitSeconds = DEFAULT_LOCK_RETRY_WAIT_SECONDS;
|
|
private int maximumStringLength;
|
|
|
|
private ThreadLocal<StringBuilder> executedStatementsThreadLocal = new ThreadLocal<StringBuilder>();
|
|
private File xmlPreSchemaOutputFile; // This must be set if there are any executed statements
|
|
|
|
public SchemaBootstrap()
|
|
{
|
|
preCreateScriptUrls = new ArrayList<String>(1);
|
|
postCreateScriptUrls = new ArrayList<String>(1);
|
|
validateUpdateScriptPatches = new ArrayList<SchemaUpgradeScriptPatch>(4);
|
|
preUpdateScriptPatches = new ArrayList<SchemaUpgradeScriptPatch>(4);
|
|
postUpdateScriptPatches = new ArrayList<SchemaUpgradeScriptPatch>(4);
|
|
maximumStringLength = -1;
|
|
}
|
|
|
|
public void setLocalSessionFactory(LocalSessionFactoryBean localSessionFactory)
|
|
{
|
|
this.localSessionFactory = localSessionFactory;
|
|
}
|
|
|
|
public LocalSessionFactoryBean getLocalSessionFactory()
|
|
{
|
|
return localSessionFactory;
|
|
}
|
|
|
|
/**
|
|
* Set this to output the full database creation script
|
|
*
|
|
* @param schemaOuputFilename the name of a file to dump the schema to, or null to ignore
|
|
*/
|
|
public void setSchemaOuputFilename(String schemaOuputFilename)
|
|
{
|
|
this.schemaOuputFilename = schemaOuputFilename;
|
|
}
|
|
|
|
/**
|
|
* Set whether to modify the schema or not. Either way, the schema will be validated.
|
|
*
|
|
* @param updateSchema true to update and validate the schema, otherwise false to just
|
|
* validate the schema. Default is <b>true</b>.
|
|
*/
|
|
public void setUpdateSchema(boolean updateSchema)
|
|
{
|
|
this.updateSchema = updateSchema;
|
|
}
|
|
|
|
/**
|
|
* Set whether this component should terminate the bootstrap process after running all the
|
|
* usual checks and scripts. This has the additional effect of dumping a final schema
|
|
* structure file just before exiting.
|
|
* <p>
|
|
* <b>WARNING: </b>USE FOR DEBUG AND UPGRADE TESTING ONLY
|
|
*
|
|
* @param stopAfterSchemaBootstrap <tt>true</tt> to terminate (with exception) after
|
|
* running all the usual schema updates and checks.
|
|
*/
|
|
public void setStopAfterSchemaBootstrap(boolean stopAfterSchemaBootstrap)
|
|
{
|
|
this.stopAfterSchemaBootstrap = stopAfterSchemaBootstrap;
|
|
}
|
|
|
|
/**
|
|
* Set the scripts that must be executed <b>before</b> the schema has been created.
|
|
*
|
|
* @param postCreateScriptUrls file URLs
|
|
*
|
|
* @see #PLACEHOLDER_SCRIPT_DIALECT
|
|
*/
|
|
public void setPreCreateScriptUrls(List<String> preUpdateScriptUrls)
|
|
{
|
|
this.preCreateScriptUrls = preUpdateScriptUrls;
|
|
}
|
|
|
|
/**
|
|
* Set the scripts that must be executed <b>after</b> the schema has been created.
|
|
*
|
|
* @param postCreateScriptUrls file URLs
|
|
*
|
|
* @see #PLACEHOLDER_SCRIPT_DIALECT
|
|
*/
|
|
public void setPostCreateScriptUrls(List<String> postUpdateScriptUrls)
|
|
{
|
|
this.postCreateScriptUrls = postUpdateScriptUrls;
|
|
}
|
|
|
|
/**
|
|
* Set the schema script patches that must have been applied. These will not be
|
|
* applied to the database. These can be used where the script <u>cannot</u> be
|
|
* applied automatically or where a particular upgrade path is no longer supported.
|
|
* For example, at version 3.0, the upgrade scripts for version 1.4 may be considered
|
|
* unsupported - this doesn't prevent the manual application of the scripts, though.
|
|
*
|
|
* @param scriptPatches a list of schema patches to check
|
|
*/
|
|
public void setValidateUpdateScriptPatches(List<SchemaUpgradeScriptPatch> scriptPatches)
|
|
{
|
|
this.validateUpdateScriptPatches = scriptPatches;
|
|
}
|
|
|
|
/**
|
|
* Set the schema script patches that may be applied prior to the auto-update process.
|
|
*
|
|
* @param scriptPatches a list of schema patches to check
|
|
*/
|
|
public void setPreUpdateScriptPatches(List<SchemaUpgradeScriptPatch> scriptPatches)
|
|
{
|
|
this.preUpdateScriptPatches = scriptPatches;
|
|
}
|
|
|
|
/**
|
|
* Set the schema script patches that may be applied after the auto-update process.
|
|
*
|
|
* @param postUpdateScriptPatches a list of schema patches to check
|
|
*/
|
|
public void setPostUpdateScriptPatches(List<SchemaUpgradeScriptPatch> scriptPatches)
|
|
{
|
|
this.postUpdateScriptPatches = scriptPatches;
|
|
}
|
|
|
|
/**
|
|
* Set the number times that the DB must be checked for the presence of the table
|
|
* indicating that a schema change is in progress.
|
|
*
|
|
* @param schemaUpdateLockRetryCount the number of times to retry (default 24)
|
|
*/
|
|
public void setSchemaUpdateLockRetryCount(int schemaUpdateLockRetryCount)
|
|
{
|
|
this.schemaUpdateLockRetryCount = schemaUpdateLockRetryCount;
|
|
}
|
|
|
|
/**
|
|
* Set the wait time (seconds) between checks for the schema update lock.
|
|
*
|
|
* @param schemaUpdateLockRetryWaitSeconds the number of seconds between checks (default 5 seconds)
|
|
*/
|
|
public void setSchemaUpdateLockRetryWaitSeconds(int schemaUpdateLockRetryWaitSeconds)
|
|
{
|
|
this.schemaUpdateLockRetryWaitSeconds = schemaUpdateLockRetryWaitSeconds;
|
|
}
|
|
|
|
/**
|
|
* Optionally override the system's default maximum string length. Some databases have
|
|
* limitations on how long the <b>string_value</b> columns can be while other do not.
|
|
* Some parts of the persistence have alternatives when the string values exceed this
|
|
* length while others do not. Either way, it is possible to adjust the text column sizes
|
|
* and adjust this value manually to override the default associated with the database
|
|
* being used.
|
|
* <p>
|
|
* The system - as of V2.1.2 - will attempt to adjust the maximum string length size
|
|
* automatically and therefore this method is not normally required. But it is possible
|
|
* to manually override the value if, for example, the system doesn't guess the correct
|
|
* maximum length or if the dialect is not explicitly catered for.
|
|
* <p>
|
|
* All negative or zero values are ignored and the system defaults to its best guess based
|
|
* on the dialect being used.
|
|
*
|
|
* @param maximumStringLength the maximum length of the <b>string_value</b> columns
|
|
*/
|
|
public void setMaximumStringLength(int maximumStringLength)
|
|
{
|
|
if (maximumStringLength > 0)
|
|
{
|
|
this.maximumStringLength = maximumStringLength;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Get the limit for the hibernate executions queue
|
|
* @return
|
|
*/
|
|
public int getHibernateMaxExecutions()
|
|
{
|
|
return ActionQueue.getMAX_EXECUTIONS_SIZE();
|
|
}
|
|
|
|
/**
|
|
* Set the limit for the hibernate executions queue
|
|
* Less than zero always uses event amalgamation
|
|
* @param hibernateMaxExecutions
|
|
*/
|
|
public void setHibernateMaxExecutions(int hibernateMaxExecutions)
|
|
{
|
|
ActionQueue.setMAX_EXECUTIONS_SIZE(hibernateMaxExecutions);
|
|
}
|
|
|
|
/**
|
|
* Helper method to generate a schema creation SQL script from the given Hibernate
|
|
* configuration.
|
|
*/
|
|
private static void dumpSchemaCreate(Configuration cfg, File schemaOutputFile)
|
|
{
|
|
// if the file exists, delete it
|
|
if (schemaOutputFile.exists())
|
|
{
|
|
schemaOutputFile.delete();
|
|
}
|
|
SchemaExport schemaExport = new SchemaExport(cfg)
|
|
.setFormat(true)
|
|
.setHaltOnError(true)
|
|
.setOutputFile(schemaOutputFile.getAbsolutePath())
|
|
.setDelimiter(";");
|
|
schemaExport.execute(false, false, false, true);
|
|
}
|
|
|
|
private SessionFactory getSessionFactory()
|
|
{
|
|
return (SessionFactory) localSessionFactory.getObject();
|
|
}
|
|
|
|
private static class NoSchemaException extends Exception
|
|
{
|
|
private static final long serialVersionUID = 5574280159910824660L;
|
|
}
|
|
|
|
/**
|
|
* Used to indicate a forced stop of the bootstrap.
|
|
*
|
|
* @see SchemaBootstrap#setStopAfterSchemaBootstrap(boolean)
|
|
*
|
|
* @author Derek Hulley
|
|
* @since 3.1.1
|
|
*/
|
|
private static class BootstrapStopException extends RuntimeException
|
|
{
|
|
private static final long serialVersionUID = 4250016675538442181L;
|
|
private BootstrapStopException()
|
|
{
|
|
super(I18NUtil.getMessage(ERR_FORCED_STOP));
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Count applied patches. This fails if multiple applied patch tables are found,
|
|
* which normally indicates that the schema view needs to be limited.
|
|
*
|
|
* @param cfg The Hibernate config
|
|
* @param connection a valid database connection
|
|
* @return Returns the number of applied patches
|
|
* @throws NoSchemaException if the table of applied patches can't be found
|
|
*/
|
|
private int countAppliedPatches(Configuration cfg, Connection connection) throws Exception
|
|
{
|
|
String defaultSchema = cfg.getProperty("hibernate.default_schema");
|
|
if (defaultSchema != null && defaultSchema.length() == 0)
|
|
{
|
|
defaultSchema = null;
|
|
}
|
|
String defaultCatalog = cfg.getProperty("hibernate.default_catalog");
|
|
if (defaultCatalog != null && defaultCatalog.length() == 0)
|
|
{
|
|
defaultCatalog = null;
|
|
}
|
|
DatabaseMetaData dbMetadata = connection.getMetaData();
|
|
|
|
ResultSet tableRs = dbMetadata.getTables(defaultCatalog, defaultSchema, "%", null);
|
|
boolean newPatchTable = false;
|
|
boolean oldPatchTable = false;
|
|
try
|
|
{
|
|
boolean multipleSchemas = false;
|
|
while (tableRs.next())
|
|
{
|
|
String tableName = tableRs.getString("TABLE_NAME");
|
|
if (tableName.equalsIgnoreCase("applied_patch"))
|
|
{
|
|
if (oldPatchTable || newPatchTable)
|
|
{
|
|
// Found earlier
|
|
multipleSchemas = true;
|
|
}
|
|
oldPatchTable = true;
|
|
}
|
|
else if (tableName.equalsIgnoreCase("alf_applied_patch"))
|
|
{
|
|
if (oldPatchTable || newPatchTable)
|
|
{
|
|
// Found earlier
|
|
multipleSchemas = true;
|
|
}
|
|
newPatchTable = true;
|
|
}
|
|
}
|
|
// We go through all the tables so that multiple visible schemas are detected
|
|
if (multipleSchemas)
|
|
{
|
|
throw new AlfrescoRuntimeException(ERR_MULTIPLE_SCHEMAS);
|
|
}
|
|
}
|
|
finally
|
|
{
|
|
try { tableRs.close(); } catch (Throwable e) {e.printStackTrace(); }
|
|
}
|
|
|
|
if (newPatchTable)
|
|
{
|
|
Statement stmt = connection.createStatement();
|
|
try
|
|
{
|
|
ResultSet rs = stmt.executeQuery("select count(id) from alf_applied_patch");
|
|
rs.next();
|
|
int count = rs.getInt(1);
|
|
return count;
|
|
}
|
|
catch (SQLException e)
|
|
{
|
|
// This should work at least and is probably an indication of the user viewing multiple schemas
|
|
throw new AlfrescoRuntimeException(ERR_MULTIPLE_SCHEMAS);
|
|
}
|
|
finally
|
|
{
|
|
try { stmt.close(); } catch (Throwable e) {}
|
|
}
|
|
}
|
|
else if (oldPatchTable)
|
|
{
|
|
// found the old style table name
|
|
Statement stmt = connection.createStatement();
|
|
try
|
|
{
|
|
ResultSet rs = stmt.executeQuery("select count(id) from applied_patch");
|
|
rs.next();
|
|
int count = rs.getInt(1);
|
|
return count;
|
|
}
|
|
finally
|
|
{
|
|
try { stmt.close(); } catch (Throwable e) {}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
// The applied patches table is not present
|
|
throw new NoSchemaException();
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @return Returns the name of the applied patch table, or <tt>null</tt> if the table doesn't exist
|
|
*/
|
|
private String getAppliedPatchTableName(Connection connection) throws Exception
|
|
{
|
|
Statement stmt = connection.createStatement();
|
|
try
|
|
{
|
|
stmt.executeQuery("select * from alf_applied_patch");
|
|
return "alf_applied_patch";
|
|
}
|
|
catch (Throwable e)
|
|
{
|
|
// we'll try another table name
|
|
}
|
|
finally
|
|
{
|
|
try { stmt.close(); } catch (Throwable e) {}
|
|
}
|
|
// for pre-1.4 databases, the table was named differently
|
|
stmt = connection.createStatement();
|
|
try
|
|
{
|
|
stmt.executeQuery("select * from applied_patch");
|
|
return "applied_patch";
|
|
}
|
|
catch (Throwable e)
|
|
{
|
|
// It is not there
|
|
return null;
|
|
}
|
|
finally
|
|
{
|
|
try { stmt.close(); } catch (Throwable e) {}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @return Returns the number of applied patches
|
|
*/
|
|
private boolean didPatchSucceed(Connection connection, String patchId) throws Exception
|
|
{
|
|
String patchTableName = getAppliedPatchTableName(connection);
|
|
if (patchTableName == null)
|
|
{
|
|
// Table doesn't exist, yet
|
|
return false;
|
|
}
|
|
Statement stmt = connection.createStatement();
|
|
try
|
|
{
|
|
ResultSet rs = stmt.executeQuery("select succeeded from " + patchTableName + " where id = '" + patchId + "'");
|
|
if (!rs.next())
|
|
{
|
|
return false;
|
|
}
|
|
boolean succeeded = rs.getBoolean(1);
|
|
return succeeded;
|
|
}
|
|
finally
|
|
{
|
|
try { stmt.close(); } catch (Throwable e) {}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Finds the <b>version.properties</b> file and determines the installed <b>version.schema</b>.<br>
|
|
* The only way to determine the original installed schema number is by quering the for the minimum value in
|
|
* <b>alf_applied_patch.applied_to_schema</b>. This might not work if an upgrade is attempted straight from
|
|
* Alfresco v1.0!
|
|
*
|
|
* @return the installed schema number or <tt>-1</tt> if the installation is new.
|
|
*/
|
|
private int getInstalledSchemaNumber(Connection connection) throws Exception
|
|
{
|
|
Statement stmt = connection.createStatement();
|
|
try
|
|
{
|
|
ResultSet rs = stmt.executeQuery(
|
|
"select min(applied_to_schema) from alf_applied_patch where applied_to_schema > -1");
|
|
if (!rs.next())
|
|
{
|
|
// Nothing in the table
|
|
return -1;
|
|
}
|
|
if (rs.getObject(1) == null)
|
|
{
|
|
// Nothing in the table
|
|
return -1;
|
|
}
|
|
int installedSchema = rs.getInt(1);
|
|
return installedSchema;
|
|
}
|
|
finally
|
|
{
|
|
try { stmt.close(); } catch (Throwable e) {}
|
|
}
|
|
}
|
|
|
|
private static class LockFailedException extends Exception
|
|
{
|
|
private static final long serialVersionUID = -6676398230191205456L;
|
|
}
|
|
|
|
|
|
/**
|
|
* Records that the bootstrap process has started
|
|
*/
|
|
private synchronized void setBootstrapStarted(Connection connection) throws Exception
|
|
{
|
|
// Create the marker table
|
|
Statement stmt = connection.createStatement();
|
|
try
|
|
{
|
|
stmt.executeUpdate("create table alf_bootstrap_lock (charval CHAR(1) NOT NULL)");
|
|
// Success
|
|
return;
|
|
}
|
|
catch (Throwable e)
|
|
{
|
|
// We throw a well-known exception to be handled by retrying code if required
|
|
throw new LockFailedException();
|
|
}
|
|
finally
|
|
{
|
|
try { stmt.close(); } catch (Throwable e) {}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Records that the bootstrap process has finished
|
|
*/
|
|
private void setBootstrapCompleted(Connection connection) throws Exception
|
|
{
|
|
// Create the marker table
|
|
Statement stmt = connection.createStatement();
|
|
try
|
|
{
|
|
stmt.executeUpdate("drop table alf_bootstrap_lock");
|
|
}
|
|
catch (Throwable e)
|
|
{
|
|
// Table exists
|
|
throw AlfrescoRuntimeException.create(ERR_PREVIOUS_FAILED_BOOTSTRAP);
|
|
}
|
|
finally
|
|
{
|
|
try { stmt.close(); } catch (Throwable e) {}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Builds the schema from scratch or applies the necessary patches to the schema.
|
|
*/
|
|
private boolean updateSchema(Configuration cfg, Session session, Connection connection) throws Exception
|
|
{
|
|
boolean create = false;
|
|
try
|
|
{
|
|
countAppliedPatches(cfg, connection);
|
|
}
|
|
catch (NoSchemaException e)
|
|
{
|
|
create = true;
|
|
}
|
|
// Get the dialect
|
|
final Dialect dialect = Dialect.getDialect(cfg.getProperties());
|
|
String dialectStr = dialect.getClass().getSimpleName();
|
|
|
|
if (create)
|
|
{
|
|
// execute pre-create scripts (not patches)
|
|
for (String scriptUrl : this.preCreateScriptUrls)
|
|
{
|
|
executeScriptUrl(cfg, connection, scriptUrl);
|
|
}
|
|
// Build and execute changes generated by Hibernate
|
|
File tempFile = null;
|
|
Writer writer = null;
|
|
try
|
|
{
|
|
DatabaseMetadata metadata = new DatabaseMetadata(connection, dialect);
|
|
String[] sqls = cfg.generateSchemaUpdateScript(dialect, metadata);
|
|
if (sqls.length > 0)
|
|
{
|
|
tempFile = TempFileProvider.createTempFile("AlfrescoSchema-" + dialectStr + "-Update-", ".sql");
|
|
writer = new BufferedWriter(new FileWriter(tempFile));
|
|
for (String sql : sqls)
|
|
{
|
|
writer.append(sql);
|
|
writer.append(";\n");
|
|
}
|
|
try {writer.close();} catch (Throwable e) {}
|
|
executeScriptFile(cfg, connection, tempFile, null);
|
|
}
|
|
}
|
|
finally
|
|
{
|
|
if (writer != null)
|
|
{
|
|
try {writer.close();} catch (Throwable e) {}
|
|
}
|
|
}
|
|
// execute post-create scripts (not patches)
|
|
for (String scriptUrl : this.postCreateScriptUrls)
|
|
{
|
|
executeScriptUrl(cfg, connection, scriptUrl);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
// Check for scripts that must have been run
|
|
checkSchemaPatchScripts(cfg, session, connection, validateUpdateScriptPatches, false);
|
|
// Execute any pre-auto-update scripts
|
|
checkSchemaPatchScripts(cfg, session, connection, preUpdateScriptPatches, true);
|
|
|
|
// Build and execute changes generated by Hibernate
|
|
File tempFile = null;
|
|
Writer writer = null;
|
|
try
|
|
{
|
|
DatabaseMetadata metadata = new DatabaseMetadata(connection, dialect);
|
|
String[] sqls = cfg.generateSchemaUpdateScript(dialect, metadata);
|
|
if (sqls.length > 0)
|
|
{
|
|
tempFile = TempFileProvider.createTempFile("AlfrescoSchema-" + dialectStr + "-Update-", ".sql");
|
|
writer = new BufferedWriter(new FileWriter(tempFile));
|
|
for (String sql : sqls)
|
|
{
|
|
writer.append(sql);
|
|
writer.append(";\n");
|
|
}
|
|
}
|
|
}
|
|
finally
|
|
{
|
|
if (writer != null)
|
|
{
|
|
try {writer.close();} catch (Throwable e) {}
|
|
}
|
|
}
|
|
// execute if there were changes raised by Hibernate
|
|
if (tempFile != null)
|
|
{
|
|
executeScriptFile(cfg, connection, tempFile, null);
|
|
}
|
|
|
|
// Execute any post-auto-update scripts
|
|
checkSchemaPatchScripts(cfg, session, connection, postUpdateScriptPatches, true);
|
|
}
|
|
|
|
return create;
|
|
}
|
|
|
|
/**
|
|
* Check that the necessary scripts have been executed against the database
|
|
*/
|
|
private void checkSchemaPatchScripts(
|
|
Configuration cfg,
|
|
Session session,
|
|
Connection connection,
|
|
List<SchemaUpgradeScriptPatch> scriptPatches,
|
|
boolean apply) throws Exception
|
|
{
|
|
// first check if there have been any applied patches
|
|
int appliedPatchCount = countAppliedPatches(cfg, connection);
|
|
if (appliedPatchCount == 0)
|
|
{
|
|
// This is a new schema, so upgrade scripts are irrelevant
|
|
// and patches will not have been applied yet
|
|
return;
|
|
}
|
|
// Retrieve the first installed schema number
|
|
int installedSchema = getInstalledSchemaNumber(connection);
|
|
|
|
nextPatch:
|
|
for (SchemaUpgradeScriptPatch patch : scriptPatches)
|
|
{
|
|
final String patchId = patch.getId();
|
|
final String scriptUrl = patch.getScriptUrl();
|
|
|
|
// Check if any of the alternative patches were executed
|
|
List<Patch> alternatives = patch.getAlternatives();
|
|
for (Patch alternativePatch : alternatives)
|
|
{
|
|
String alternativePatchId = alternativePatch.getId();
|
|
boolean alternativeSucceeded = didPatchSucceed(connection, alternativePatchId);
|
|
if (alternativeSucceeded)
|
|
{
|
|
continue nextPatch;
|
|
}
|
|
}
|
|
|
|
// check if the script was successfully executed
|
|
boolean wasSuccessfullyApplied = didPatchSucceed(connection, patchId);
|
|
if (wasSuccessfullyApplied)
|
|
{
|
|
// Either the patch was executed before or the system was bootstrapped
|
|
// with the patch bean present.
|
|
continue;
|
|
}
|
|
else if (!patch.applies(installedSchema))
|
|
{
|
|
// Patch does not apply to the installed schema number
|
|
continue;
|
|
}
|
|
else if (!apply)
|
|
{
|
|
// the script was not run and may not be run automatically
|
|
throw AlfrescoRuntimeException.create(ERR_SCRIPT_NOT_RUN, scriptUrl);
|
|
}
|
|
// it wasn't run and it can be run now
|
|
executeScriptUrl(cfg, connection, scriptUrl);
|
|
}
|
|
}
|
|
|
|
private void executeScriptUrl(Configuration cfg, Connection connection, String scriptUrl) throws Exception
|
|
{
|
|
Dialect dialect = Dialect.getDialect(cfg.getProperties());
|
|
String dialectStr = dialect.getClass().getSimpleName();
|
|
InputStream scriptInputStream = getScriptInputStream(dialect.getClass(), scriptUrl);
|
|
// check that it exists
|
|
if (scriptInputStream == null)
|
|
{
|
|
throw AlfrescoRuntimeException.create(ERR_SCRIPT_NOT_FOUND, scriptUrl);
|
|
}
|
|
// write the script to a temp location for future and failure reference
|
|
File tempFile = null;
|
|
try
|
|
{
|
|
tempFile = TempFileProvider.createTempFile("AlfrescoSchema-" + dialectStr + "-Update-", ".sql");
|
|
ContentWriter writer = new FileContentWriter(tempFile);
|
|
writer.putContent(scriptInputStream);
|
|
}
|
|
finally
|
|
{
|
|
try { scriptInputStream.close(); } catch (Throwable e) {} // usually a duplicate close
|
|
}
|
|
// now execute it
|
|
String dialectScriptUrl = scriptUrl.replaceAll(PLACEHOLDER_SCRIPT_DIALECT, dialect.getClass().getName());
|
|
// Replace the script placeholders
|
|
executeScriptFile(cfg, connection, tempFile, dialectScriptUrl);
|
|
}
|
|
|
|
/**
|
|
* Replaces the dialect placeholder in the script URL and attempts to find a file for
|
|
* it. If not found, the dialect hierarchy will be walked until a compatible script is
|
|
* found. This makes it possible to have scripts that are generic to all dialects.
|
|
*
|
|
* @return Returns an input stream onto the script, otherwise null
|
|
*/
|
|
private InputStream getScriptInputStream(Class dialectClazz, String scriptUrl) throws Exception
|
|
{
|
|
// replace the dialect placeholder
|
|
String dialectScriptUrl = scriptUrl.replaceAll(PLACEHOLDER_SCRIPT_DIALECT, dialectClazz.getName());
|
|
// get a handle on the resource
|
|
Resource resource = rpr.getResource(dialectScriptUrl);
|
|
if (!resource.exists())
|
|
{
|
|
// it wasn't found. Get the superclass of the dialect and try again
|
|
Class superClazz = dialectClazz.getSuperclass();
|
|
if (Dialect.class.isAssignableFrom(superClazz))
|
|
{
|
|
// we still have a Dialect - try again
|
|
return getScriptInputStream(superClazz, scriptUrl);
|
|
}
|
|
else
|
|
{
|
|
// we have exhausted all options
|
|
return null;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
// we have a handle to it
|
|
return resource.getInputStream();
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @param cfg the Hibernate configuration
|
|
* @param connection the DB connection to use
|
|
* @param scriptFile the file containing the statements
|
|
* @param scriptUrl the URL of the script to report. If this is null, the script
|
|
* is assumed to have been auto-generated.
|
|
*/
|
|
private void executeScriptFile(
|
|
Configuration cfg,
|
|
Connection connection,
|
|
File scriptFile,
|
|
String scriptUrl) throws Exception
|
|
{
|
|
final Dialect dialect = Dialect.getDialect(cfg.getProperties());
|
|
|
|
StringBuilder executedStatements = executedStatementsThreadLocal.get();
|
|
if (executedStatements == null)
|
|
{
|
|
// Dump the normalized, pre-upgrade Alfresco schema. We keep the file for later reporting.
|
|
xmlPreSchemaOutputFile = dumpSchema(
|
|
connection,
|
|
this.dialect,
|
|
TempFileProvider.createTempFile(
|
|
"AlfrescoSchema-" + this.dialect.getClass().getSimpleName() + "-",
|
|
"-Startup.xml").getPath(),
|
|
"Failed to dump normalized, pre-upgrade schema to file.");
|
|
|
|
// There is no lock at this stage. This process can fall out if the lock can't be applied.
|
|
setBootstrapStarted(connection);
|
|
executedStatements = new StringBuilder(8094);
|
|
executedStatementsThreadLocal.set(executedStatements);
|
|
}
|
|
|
|
if (scriptUrl == null)
|
|
{
|
|
LogUtil.info(logger, MSG_EXECUTING_GENERATED_SCRIPT, scriptFile);
|
|
}
|
|
else
|
|
{
|
|
LogUtil.info(logger, MSG_EXECUTING_COPIED_SCRIPT, scriptFile, scriptUrl);
|
|
}
|
|
|
|
InputStream scriptInputStream = new FileInputStream(scriptFile);
|
|
BufferedReader reader = new BufferedReader(new InputStreamReader(scriptInputStream, "UTF-8"));
|
|
try
|
|
{
|
|
int line = 0;
|
|
// loop through all statements
|
|
StringBuilder sb = new StringBuilder(1024);
|
|
String fetchVarName = null;
|
|
String fetchColumnName = null;
|
|
Map<String, Object> varAssignments = new HashMap<String, Object>(13);
|
|
// Special variable assignments:
|
|
if (dialect instanceof PostgreSQLDialect)
|
|
{
|
|
// Needs 1/0 for true/false
|
|
varAssignments.put("true", "true");
|
|
varAssignments.put("false", "false");
|
|
varAssignments.put("TRUE", "TRUE");
|
|
varAssignments.put("FALSE", "FALSE");
|
|
}
|
|
else
|
|
{
|
|
// Needs true/false as strings
|
|
varAssignments.put("true", "1");
|
|
varAssignments.put("false", "0");
|
|
varAssignments.put("TRUE", "1");
|
|
varAssignments.put("FALSE", "0");
|
|
}
|
|
|
|
while(true)
|
|
{
|
|
String sqlOriginal = reader.readLine();
|
|
line++;
|
|
|
|
if (sqlOriginal == null)
|
|
{
|
|
// nothing left in the file
|
|
break;
|
|
}
|
|
|
|
// trim it
|
|
String sql = sqlOriginal.trim();
|
|
// Check for variable assignment
|
|
if (sql.startsWith("--ASSIGN:"))
|
|
{
|
|
if (sb.length() > 0)
|
|
{
|
|
// This can only be set before a new SQL statement
|
|
throw AlfrescoRuntimeException.create(ERR_STATEMENT_VAR_ASSIGNMENT_BEFORE_SQL, (line - 1), scriptUrl);
|
|
}
|
|
String assignStr = sql.substring(9, sql.length());
|
|
String[] assigns = assignStr.split("=");
|
|
if (assigns.length != 2 || assigns[0].length() == 0 || assigns[1].length() == 0)
|
|
{
|
|
throw AlfrescoRuntimeException.create(ERR_STATEMENT_VAR_ASSIGNMENT_FORMAT, (line - 1), scriptUrl);
|
|
}
|
|
fetchVarName = assigns[0];
|
|
fetchColumnName = assigns[1];
|
|
continue;
|
|
}
|
|
// Allow transaction delineation
|
|
else if (sql.startsWith("--BEGIN TXN"))
|
|
{
|
|
connection.setAutoCommit(false);
|
|
}
|
|
else if (sql.startsWith("--END TXN"))
|
|
{
|
|
connection.commit();
|
|
connection.setAutoCommit(true);
|
|
}
|
|
|
|
// Check for comments
|
|
if (sql.length() == 0 ||
|
|
sql.startsWith( "--" ) ||
|
|
sql.startsWith( "//" ) ||
|
|
sql.startsWith( "/*" ) )
|
|
{
|
|
if (sb.length() > 0)
|
|
{
|
|
// we have an unterminated statement
|
|
throw AlfrescoRuntimeException.create(ERR_STATEMENT_TERMINATOR, (line - 1), scriptUrl);
|
|
}
|
|
// there has not been anything to execute - it's just a comment line
|
|
continue;
|
|
}
|
|
// have we reached the end of a statement?
|
|
boolean execute = false;
|
|
boolean optional = false;
|
|
if (sql.endsWith(";"))
|
|
{
|
|
sql = sql.substring(0, sql.length() - 1);
|
|
execute = true;
|
|
optional = false;
|
|
}
|
|
else if (sql.endsWith("(optional)") || sql.endsWith("(OPTIONAL)"))
|
|
{
|
|
// Get the end of statement
|
|
int endIndex = sql.lastIndexOf(';');
|
|
if (endIndex > -1)
|
|
{
|
|
sql = sql.substring(0, endIndex);
|
|
execute = true;
|
|
optional = true;
|
|
}
|
|
else
|
|
{
|
|
// Ends with "(optional)" but there is no semi-colon.
|
|
// Just take it at face value and probably fail.
|
|
}
|
|
}
|
|
// Add newline
|
|
if (sb.length() > 0)
|
|
{
|
|
sb.append("\n");
|
|
}
|
|
// Add leading whitespace for formatting
|
|
int whitespaceCount = sqlOriginal.indexOf(sql);
|
|
for (int i = 0; i < whitespaceCount; i++)
|
|
{
|
|
sb.append(" ");
|
|
}
|
|
// append to the statement being built up
|
|
sb.append(sql);
|
|
// execute, if required
|
|
if (execute)
|
|
{
|
|
sql = sb.toString();
|
|
|
|
// Perform variable replacement using the ${var} format
|
|
for (Map.Entry<String, Object> entry : varAssignments.entrySet())
|
|
{
|
|
String var = entry.getKey();
|
|
Object val = entry.getValue();
|
|
sql = sql.replaceAll("\\$\\{" + var + "\\}", val.toString());
|
|
}
|
|
|
|
// Handle the 0/1 values that PostgreSQL doesn't translate to TRUE
|
|
if (this.dialect != null && this.dialect instanceof PostgreSQLDialect)
|
|
{
|
|
sql = sql.replaceAll("\\$\\{TRUE\\}", "TRUE");
|
|
}
|
|
else
|
|
{
|
|
sql = sql.replaceAll("\\$\\{TRUE\\}", "1");
|
|
}
|
|
|
|
Object fetchedVal = executeStatement(connection, sql, fetchColumnName, optional, line, scriptFile);
|
|
if (fetchVarName != null && fetchColumnName != null)
|
|
{
|
|
varAssignments.put(fetchVarName, fetchedVal);
|
|
}
|
|
sb = new StringBuilder(1024);
|
|
fetchVarName = null;
|
|
fetchColumnName = null;
|
|
}
|
|
}
|
|
}
|
|
finally
|
|
{
|
|
try { reader.close(); } catch (Throwable e) {}
|
|
try { scriptInputStream.close(); } catch (Throwable e) {}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Execute the given SQL statement, absorbing exceptions that we expect during
|
|
* schema creation or upgrade.
|
|
*
|
|
* @param fetchColumnName the name of the column value to return
|
|
*/
|
|
private Object executeStatement(
|
|
Connection connection,
|
|
String sql,
|
|
String fetchColumnName,
|
|
boolean optional,
|
|
int line,
|
|
File file) throws Exception
|
|
{
|
|
StringBuilder executedStatements = executedStatementsThreadLocal.get();
|
|
if (executedStatements == null)
|
|
{
|
|
throw new IllegalArgumentException("The executedStatementsThreadLocal must be populated");
|
|
}
|
|
|
|
Statement stmt = connection.createStatement();
|
|
Object ret = null;
|
|
try
|
|
{
|
|
if (logger.isDebugEnabled())
|
|
{
|
|
LogUtil.debug(logger, MSG_EXECUTING_STATEMENT, sql);
|
|
}
|
|
boolean haveResults = stmt.execute(sql);
|
|
// Record the statement
|
|
executedStatements.append(sql).append(";\n\n");
|
|
if (haveResults && fetchColumnName != null)
|
|
{
|
|
ResultSet rs = stmt.getResultSet();
|
|
if (rs.next())
|
|
{
|
|
// Get the result value
|
|
ret = rs.getObject(fetchColumnName);
|
|
}
|
|
}
|
|
}
|
|
catch (SQLException e)
|
|
{
|
|
if (optional)
|
|
{
|
|
// it was marked as optional, so we just ignore it
|
|
LogUtil.debug(logger, MSG_OPTIONAL_STATEMENT_FAILED, sql, e.getMessage(), file.getAbsolutePath(), line);
|
|
}
|
|
else
|
|
{
|
|
LogUtil.error(logger, ERR_STATEMENT_FAILED, sql, e.getMessage(), file.getAbsolutePath(), line);
|
|
throw e;
|
|
}
|
|
}
|
|
finally
|
|
{
|
|
try { stmt.close(); } catch (Throwable e) {}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Performs dialect-specific checking. This includes checking for InnoDB, dumping the dialect being used
|
|
* as well as setting any runtime, dialect-specific properties.
|
|
*/
|
|
private void checkDialect(Dialect dialect)
|
|
{
|
|
Class dialectClazz = dialect.getClass();
|
|
LogUtil.info(logger, MSG_DIALECT_USED, dialectClazz.getName());
|
|
if (dialectClazz.equals(MySQLDialect.class) || dialectClazz.equals(MySQL5Dialect.class))
|
|
{
|
|
LogUtil.error(logger, ERR_DIALECT_SHOULD_USE, dialectClazz.getName(), MySQLInnoDBDialect.class.getName());
|
|
throw AlfrescoRuntimeException.create(WARN_DIALECT_UNSUPPORTED, dialectClazz.getName());
|
|
}
|
|
else if (dialectClazz.equals(HSQLDialect.class))
|
|
{
|
|
LogUtil.info(logger, WARN_DIALECT_HSQL);
|
|
}
|
|
else if (dialectClazz.equals(DerbyDialect.class))
|
|
{
|
|
LogUtil.info(logger, WARN_DIALECT_DERBY);
|
|
}
|
|
else if (dialectClazz.equals(Oracle9iDialect.class) || dialectClazz.equals(Oracle10gDialect.class))
|
|
{
|
|
LogUtil.error(logger, ERR_DIALECT_SHOULD_USE, dialectClazz.getName(), AlfrescoOracle9Dialect.class.getName());
|
|
throw AlfrescoRuntimeException.create(WARN_DIALECT_UNSUPPORTED, dialectClazz.getName());
|
|
}
|
|
else if (dialectClazz.equals(OracleDialect.class) || dialectClazz.equals(Oracle9Dialect.class))
|
|
{
|
|
LogUtil.error(logger, ERR_DIALECT_SHOULD_USE, dialectClazz.getName(), AlfrescoOracle9Dialect.class.getName());
|
|
throw AlfrescoRuntimeException.create(WARN_DIALECT_UNSUPPORTED, dialectClazz.getName());
|
|
}
|
|
|
|
int maxStringLength = SchemaBootstrap.DEFAULT_MAX_STRING_LENGTH;
|
|
int serializableType = SerializableTypeHandlerCallback.getSerializableType();
|
|
// Adjust the maximum allowable String length according to the dialect
|
|
if (dialect instanceof AlfrescoSQLServerDialect)
|
|
{
|
|
// string_value nvarchar(1024) null,
|
|
// serializable_value image null,
|
|
maxStringLength = SchemaBootstrap.DEFAULT_MAX_STRING_LENGTH;
|
|
}
|
|
else if (dialect instanceof AlfrescoSybaseAnywhereDialect)
|
|
{
|
|
// string_value text null,
|
|
// serializable_value varbinary(8192) null,
|
|
maxStringLength = Integer.MAX_VALUE;
|
|
}
|
|
else if (dialect instanceof DB2Dialect)
|
|
{
|
|
// string_value varchar(1024),
|
|
// serializable_value varchar(8192) for bit data,
|
|
maxStringLength = SchemaBootstrap.DEFAULT_MAX_STRING_LENGTH;
|
|
serializableType = Types.BLOB;
|
|
}
|
|
else if (dialect instanceof HSQLDialect)
|
|
{
|
|
// string_value varchar(1024),
|
|
// serializable_value varbinary(8192),
|
|
maxStringLength = SchemaBootstrap.DEFAULT_MAX_STRING_LENGTH;
|
|
}
|
|
else if (dialect instanceof MySQLInnoDBDialect)
|
|
{
|
|
// string_value text,
|
|
// serializable_value blob,
|
|
maxStringLength = Integer.MAX_VALUE;
|
|
}
|
|
else if (dialect instanceof AlfrescoOracle9Dialect)
|
|
{
|
|
// string_value varchar2(1024 char),
|
|
// serializable_value blob,
|
|
maxStringLength = SchemaBootstrap.DEFAULT_MAX_STRING_LENGTH;
|
|
}
|
|
else if (dialect instanceof PostgreSQLDialect)
|
|
{
|
|
// string_value varchar(1024),
|
|
// serializable_value bytea,
|
|
maxStringLength = SchemaBootstrap.DEFAULT_MAX_STRING_LENGTH;
|
|
}
|
|
SchemaBootstrap.setMaxStringLength(maxStringLength);
|
|
SerializableTypeHandlerCallback.setSerializableType(serializableType);
|
|
|
|
// Now override the maximum string length if it was set directly
|
|
if (maximumStringLength > 0)
|
|
{
|
|
SchemaBootstrap.setMaxStringLength(maximumStringLength);
|
|
}
|
|
}
|
|
|
|
@Override
|
|
protected synchronized void onBootstrap(ApplicationEvent event)
|
|
{
|
|
// Use the application context to load resources
|
|
rpr = (ApplicationContext)event.getSource();
|
|
|
|
// do everything in a transaction
|
|
Session session = getSessionFactory().openSession();
|
|
Connection connection = null;
|
|
try
|
|
{
|
|
// make sure that we AUTO-COMMIT
|
|
connection = session.connection();
|
|
connection.setAutoCommit(true);
|
|
|
|
Configuration cfg = localSessionFactory.getConfiguration();
|
|
|
|
// Check and dump the dialect being used
|
|
checkDialect(this.dialect);
|
|
|
|
// Ensure that our static connection provider is used
|
|
String defaultConnectionProviderFactoryClass = cfg.getProperty(Environment.CONNECTION_PROVIDER);
|
|
cfg.setProperty(Environment.CONNECTION_PROVIDER, SchemaBootstrapConnectionProvider.class.getName());
|
|
SchemaBootstrapConnectionProvider.setBootstrapConnection(connection);
|
|
|
|
// Update the schema, if required.
|
|
if (updateSchema)
|
|
{
|
|
// Retries are required here as the DB lock will be applied lazily upon first statement execution.
|
|
// So if the schema is up to date (no statements executed) then the LockFailException cannot be
|
|
// thrown. If it is thrown, the the update needs to be rerun as it will probably generate no SQL
|
|
// statements the second time around.
|
|
boolean updatedSchema = false;
|
|
boolean createdSchema = false;
|
|
|
|
for (int i = 0; i < schemaUpdateLockRetryCount; i++)
|
|
{
|
|
try
|
|
{
|
|
createdSchema = updateSchema(cfg, session, connection);
|
|
updatedSchema = true;
|
|
break;
|
|
}
|
|
catch (LockFailedException e)
|
|
{
|
|
try { this.wait(schemaUpdateLockRetryWaitSeconds * 1000L); } catch (InterruptedException ee) {}
|
|
}
|
|
}
|
|
|
|
if (!updatedSchema)
|
|
{
|
|
// The retries were exceeded
|
|
throw new AlfrescoRuntimeException(ERR_PREVIOUS_FAILED_BOOTSTRAP);
|
|
}
|
|
|
|
// Copy the executed statements to the output file
|
|
File schemaOutputFile = null;
|
|
if (schemaOuputFilename != null)
|
|
{
|
|
schemaOutputFile = new File(schemaOuputFilename);
|
|
}
|
|
else
|
|
{
|
|
schemaOutputFile = TempFileProvider.createTempFile(
|
|
"AlfrescoSchema-" + this.dialect.getClass().getSimpleName() + "-All_Statements-",
|
|
".sql");
|
|
}
|
|
|
|
StringBuilder executedStatements = executedStatementsThreadLocal.get();
|
|
if (executedStatements == null)
|
|
{
|
|
LogUtil.info(logger, MSG_NO_CHANGES);
|
|
}
|
|
else
|
|
{
|
|
FileContentWriter writer = new FileContentWriter(schemaOutputFile);
|
|
writer.setEncoding("UTF-8");
|
|
String executedStatementsStr = executedStatements.toString();
|
|
writer.putContent(executedStatementsStr);
|
|
LogUtil.info(logger, MSG_ALL_STATEMENTS, schemaOutputFile.getPath());
|
|
}
|
|
|
|
if (! createdSchema)
|
|
{
|
|
// verify that all patches have been applied correctly
|
|
checkSchemaPatchScripts(cfg, session, connection, validateUpdateScriptPatches, false); // check scripts
|
|
checkSchemaPatchScripts(cfg, session, connection, preUpdateScriptPatches, false); // check scripts
|
|
checkSchemaPatchScripts(cfg, session, connection, postUpdateScriptPatches, false); // check scripts
|
|
}
|
|
|
|
if (executedStatements != null)
|
|
{
|
|
// Remove the flag indicating a running bootstrap
|
|
setBootstrapCompleted(connection);
|
|
}
|
|
|
|
// Report normalized dumps
|
|
if (executedStatements != null)
|
|
{
|
|
// Dump the normalized, post-upgrade Alfresco schema.
|
|
File xmlPostSchemaOutputFile = dumpSchema(
|
|
connection,
|
|
this.dialect,
|
|
TempFileProvider.createTempFile(
|
|
"AlfrescoSchema-" + this.dialect.getClass().getSimpleName() + "-",
|
|
".xml").getPath(),
|
|
"Failed to dump normalized, post-upgrade schema to file.");
|
|
|
|
if (createdSchema)
|
|
{
|
|
// This is a new schema
|
|
if (xmlPostSchemaOutputFile != null)
|
|
{
|
|
LogUtil.info(logger, MSG_NORMALIZED_SCHEMA, xmlPostSchemaOutputFile.getPath());
|
|
}
|
|
}
|
|
else
|
|
{
|
|
// We upgraded, so have to report pre- and post- schema dumps
|
|
if (xmlPreSchemaOutputFile != null)
|
|
{
|
|
LogUtil.info(logger, MSG_NORMALIZED_SCHEMA_PRE, xmlPreSchemaOutputFile.getPath());
|
|
}
|
|
if (xmlPostSchemaOutputFile != null)
|
|
{
|
|
LogUtil.info(logger, MSG_NORMALIZED_SCHEMA_POST, xmlPostSchemaOutputFile.getPath());
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
LogUtil.info(logger, MSG_BYPASSING_SCHEMA_UPDATE);
|
|
}
|
|
|
|
if (stopAfterSchemaBootstrap)
|
|
{
|
|
// We have been forced to stop, so we do one last dump of the schema and throw an exception to
|
|
// escape further startup procedures
|
|
File xmlStopSchemaOutputFile = dumpSchema(
|
|
connection,
|
|
this.dialect,
|
|
TempFileProvider.createTempFile(
|
|
"AlfrescoSchema-" + this.dialect.getClass().getSimpleName() + "-",
|
|
"-ForcedExit.xml").getPath(),
|
|
"Failed to dump normalized, post-upgrade, forced-exit schema to file.");
|
|
if (xmlStopSchemaOutputFile != null)
|
|
{
|
|
LogUtil.info(logger, MSG_NORMALIZED_SCHEMA, xmlStopSchemaOutputFile);
|
|
}
|
|
LogUtil.error(logger, ERR_FORCED_STOP);
|
|
throw new BootstrapStopException();
|
|
}
|
|
|
|
// Reset the configuration
|
|
cfg.setProperty(Environment.CONNECTION_PROVIDER, defaultConnectionProviderFactoryClass);
|
|
|
|
// all done successfully
|
|
((ApplicationContext) event.getSource()).publishEvent(new SchemaAvailableEvent(this));
|
|
}
|
|
catch (BootstrapStopException e)
|
|
{
|
|
// We just let this out
|
|
throw e;
|
|
}
|
|
catch (Throwable e)
|
|
{
|
|
LogUtil.error(logger, e, ERR_UPDATE_FAILED);
|
|
if (updateSchema)
|
|
{
|
|
throw new AlfrescoRuntimeException(ERR_UPDATE_FAILED, e);
|
|
}
|
|
else
|
|
{
|
|
throw new AlfrescoRuntimeException(ERR_VALIDATION_FAILED, e);
|
|
}
|
|
}
|
|
finally
|
|
{
|
|
try
|
|
{
|
|
if (connection != null)
|
|
{
|
|
connection.close();
|
|
}
|
|
}
|
|
catch (Throwable e)
|
|
{
|
|
logger.warn("Error closing DB connection: " + e.getMessage());
|
|
}
|
|
// Remove the connection reference from the threadlocal boostrap
|
|
SchemaBootstrapConnectionProvider.setBootstrapConnection(null);
|
|
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @return Returns the file that was written to or <tt>null</tt> if it failed
|
|
*/
|
|
private File dumpSchema(Connection connection, Dialect dialect, String fileName, String err)
|
|
{
|
|
File xmlSchemaOutputFile = new File(fileName);
|
|
try
|
|
{
|
|
Main xmlSchemaOutputMain = new Main(connection, dialect);
|
|
xmlSchemaOutputMain.execute(xmlSchemaOutputFile);
|
|
}
|
|
catch (Throwable e)
|
|
{
|
|
xmlSchemaOutputFile = null;
|
|
// Don't fail the upgrade on this
|
|
if (logger.isDebugEnabled())
|
|
{
|
|
logger.debug(err, e);
|
|
}
|
|
else
|
|
{
|
|
logger.error(err + " Error: " + e.getMessage());
|
|
}
|
|
}
|
|
return xmlSchemaOutputFile;
|
|
}
|
|
|
|
@Override
|
|
protected void onShutdown(ApplicationEvent event)
|
|
{
|
|
// Shut down DB, if required
|
|
Class<?> dialectClazz = this.dialect.getClass();
|
|
if (dialectClazz.equals(DerbyDialect.class))
|
|
{
|
|
try
|
|
{
|
|
DriverManager.getConnection("jdbc:derby:;shutdown=true");
|
|
}
|
|
// Derby shutdown always triggers an exception, even when clean
|
|
catch (Throwable e)
|
|
{
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* This is a workaround for the odd Spring-Hibernate interaction during configuration.
|
|
* The Hibernate code assumes that schema scripts will be generated during context
|
|
* initialization. We want to do it afterwards and have a little more control. Hence this class.
|
|
* <p>
|
|
* The connection that is used will not be closed or manipulated in any way. This class
|
|
* merely serves to give the connection to Hibernate.
|
|
*
|
|
* @author Derek Hulley
|
|
*/
|
|
public static class SchemaBootstrapConnectionProvider extends UserSuppliedConnectionProvider
|
|
{
|
|
private static ThreadLocal<Connection> threadLocalConnection = new ThreadLocal<Connection>();
|
|
|
|
public SchemaBootstrapConnectionProvider()
|
|
{
|
|
}
|
|
|
|
/**
|
|
* Set the connection for Hibernate to use for schema generation.
|
|
*/
|
|
public static void setBootstrapConnection(Connection connection)
|
|
{
|
|
threadLocalConnection.set(connection);
|
|
}
|
|
|
|
/**
|
|
* Unsets the connection.
|
|
*/
|
|
@Override
|
|
public void close()
|
|
{
|
|
// Leave the connection well alone, just remove it
|
|
threadLocalConnection.set(null);
|
|
}
|
|
|
|
/**
|
|
* Does nothing. The connection was given by a 3rd party and they can close it.
|
|
*/
|
|
@Override
|
|
public void closeConnection(Connection conn)
|
|
{
|
|
}
|
|
|
|
/**
|
|
* Does nothing.
|
|
*/
|
|
@Override
|
|
public void configure(Properties props) throws HibernateException
|
|
{
|
|
}
|
|
|
|
/**
|
|
* @see #setBootstrapConnection(Connection)
|
|
*/
|
|
@Override
|
|
public Connection getConnection()
|
|
{
|
|
return threadLocalConnection.get();
|
|
}
|
|
|
|
@Override
|
|
public boolean supportsAggressiveRelease()
|
|
{
|
|
return false;
|
|
}
|
|
}
|
|
|
|
private static final String DIR_SCHEMAS = "schemas";
|
|
/**
|
|
* Dump a set of creation files for all known Hibernate dialects. These will be
|
|
* dumped into the default temporary location in a subdirectory named <b>schemas</b>.
|
|
*/
|
|
public static void main(String[] args)
|
|
{
|
|
int exitCode = 0;
|
|
try
|
|
{
|
|
exitCode = dumpDialects(args);
|
|
}
|
|
catch (Throwable e)
|
|
{
|
|
LogUtil.error(logger, e, "SchemaBootstrap script dump failed");
|
|
exitCode = 1;
|
|
}
|
|
// We can exit
|
|
System.exit(exitCode);
|
|
}
|
|
|
|
private static int dumpDialects(String[] dialectClassNames)
|
|
{
|
|
if (dialectClassNames.length == 0)
|
|
{
|
|
System.out.println(
|
|
"\n" +
|
|
" ERROR: A list of fully qualified class names is required");
|
|
return 1;
|
|
}
|
|
ApplicationContext ctx = ApplicationContextHelper.getApplicationContext();
|
|
SchemaBootstrap schemaBootstrap = (SchemaBootstrap) ctx.getBean("schemaBootstrap");
|
|
LocalSessionFactoryBean localSessionFactoryBean = schemaBootstrap.getLocalSessionFactory();
|
|
Configuration configuration = localSessionFactoryBean.getConfiguration();
|
|
|
|
ServiceRegistry serviceRegistry = (ServiceRegistry) ctx.getBean(ServiceRegistry.SERVICE_REGISTRY);
|
|
DescriptorService descriptorService = serviceRegistry.getDescriptorService();
|
|
Descriptor descriptor = descriptorService.getServerDescriptor();
|
|
|
|
File tempDir = TempFileProvider.getTempDir();
|
|
File schemasDir = new File(tempDir, DIR_SCHEMAS);
|
|
if (!schemasDir.exists())
|
|
{
|
|
schemasDir.mkdir();
|
|
}
|
|
File dumpDir = new File(schemasDir, descriptor.getVersion());
|
|
if (!dumpDir.exists())
|
|
{
|
|
dumpDir.mkdir();
|
|
}
|
|
for (String dialectClassName : dialectClassNames)
|
|
{
|
|
Class dialectClazz = null;
|
|
try
|
|
{
|
|
dialectClazz = Class.forName(dialectClassName);
|
|
}
|
|
catch (ClassNotFoundException e)
|
|
{
|
|
System.out.println(
|
|
"\n" +
|
|
" ERROR: Class not found: " + dialectClassName);
|
|
continue;
|
|
}
|
|
if (!Dialect.class.isAssignableFrom(dialectClazz))
|
|
{
|
|
System.out.println(
|
|
"\n" +
|
|
" ERROR: The class name is not a valid dialect: " + dialectClassName);
|
|
continue;
|
|
}
|
|
dumpDialectScript(configuration, dialectClazz, dumpDir);
|
|
}
|
|
// Done
|
|
return 0;
|
|
}
|
|
|
|
private static void dumpDialectScript(Configuration configuration, Class dialectClazz, File directory)
|
|
{
|
|
// Set the dialect
|
|
configuration.setProperty("hibernate.dialect", dialectClazz.getName());
|
|
|
|
// First dump the dialect's schema
|
|
String filename = "default-schema-create-" + dialectClazz.getName() + ".sql";
|
|
File dumpFile = new File(directory, filename);
|
|
|
|
// Write the file
|
|
SchemaBootstrap.dumpSchemaCreate(configuration, dumpFile);
|
|
}
|
|
}
|