All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.avaje.ebean.config.ServerConfig Maven / Gradle / Ivy

There is a newer version: 8.1.1
Show newest version
package com.avaje.ebean.config;

import com.avaje.ebean.EbeanServerFactory;
import com.avaje.ebean.PersistenceContextScope;
import com.avaje.ebean.annotation.Encrypted;
import com.avaje.ebean.cache.ServerCacheManager;
import com.avaje.ebean.cache.ServerCachePlugin;
import com.avaje.ebean.config.dbplatform.DatabasePlatform;
import com.avaje.ebean.config.dbplatform.DbEncrypt;
import com.avaje.ebean.event.BeanFindController;
import com.avaje.ebean.event.BeanPersistController;
import com.avaje.ebean.event.BeanPersistListener;
import com.avaje.ebean.event.BeanPostLoad;
import com.avaje.ebean.event.BeanQueryAdapter;
import com.avaje.ebean.event.BulkTableEventListener;
import com.avaje.ebean.event.ServerConfigStartup;
import com.avaje.ebean.event.TransactionEventListener;
import com.avaje.ebean.event.changelog.ChangeLogListener;
import com.avaje.ebean.event.changelog.ChangeLogPrepare;
import com.avaje.ebean.event.changelog.ChangeLogRegister;
import com.avaje.ebean.event.readaudit.ReadAuditLogger;
import com.avaje.ebean.event.readaudit.ReadAuditPrepare;
import com.avaje.ebean.meta.MetaInfoManager;
import com.fasterxml.jackson.core.JsonFactory;
import org.avaje.datasource.DataSourceConfig;

import javax.sql.DataSource;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.Iterator;
import java.util.List;
import java.util.Map;
import java.util.Properties;
import java.util.ServiceLoader;

/**
 * The configuration used for creating a EbeanServer.
 * 

* Used to programmatically construct an EbeanServer and optionally register it * with the Ebean singleton. *

*

* If you just use Ebean without this programmatic configuration Ebean will read * the ebean.properties file and take the configuration from there. This usually * includes searching the class path and automatically registering any entity * classes and listeners etc. *

* *
{@code
 *
 * ServerConfig c = new ServerConfig();
 * c.setName("db");
 *
 * // read the ebean.properties and load
 * // those settings into this serverConfig object
 * c.loadFromProperties();
 *
 * // add any classes found in the app.data package
 * c.addPackage("com.myapp.domain");
 *
 * // register as the 'Default' server
 * c.setDefaultServer(true);
 *
 * EbeanServer server = EbeanServerFactory.create(c);
 *
 * }
* * @author emcgreal * @author rbygrave * @see EbeanServerFactory */ public class ServerConfig { /** * The EbeanServer name. */ private String name; /** * Typically configuration type objects that are passed by this ServerConfig * to plugins. For example - IgniteConfiguration passed to Ignite plugin. */ private Map serviceObject = new HashMap(); private ContainerConfig containerConfig; /** * The underlying properties that were used during configuration. */ private Properties properties; /** * The resource directory. */ private String resourceDirectory; /** * Set to true to register this EbeanServer with the Ebean singleton. */ private boolean register = true; /** * Set to true if this is the default/primary server. */ private boolean defaultServer; /** * Set this to true to disable class path search. */ private boolean disableClasspathSearch; /** * List of interesting classes such as entities, embedded, ScalarTypes, * Listeners, Finders, Controllers etc. */ private List> classes = new ArrayList>(); /** * The packages that are searched for interesting classes. Only used when * classes is empty/not explicitly specified. */ private List packages = new ArrayList(); /** * Configuration for the ElasticSearch integration. */ private DocStoreConfig docStoreConfig = new DocStoreConfig(); /** * This is used to populate @WhoCreated, @WhoModified and * support other audit features (who executed a query etc). */ private CurrentUserProvider currentUserProvider; /** * Config controlling the AutoTune behaviour. */ private AutoTuneConfig autoTuneConfig = new AutoTuneConfig(); /** * The JSON format used for DateTime types. Default to millis. */ private JsonConfig.DateTime jsonDateTime = JsonConfig.DateTime.MILLIS; /** * For writing JSON specify if null values or empty collections should be exluded. * By default all values are included. */ private JsonConfig.Include jsonInclude = JsonConfig.Include.ALL; /** * The database platform name. Used to imply a DatabasePlatform to use. */ private String databasePlatformName; /** * The database platform. */ private DatabasePlatform databasePlatform; /** * For DB's using sequences this is the number of sequence values prefetched. */ private int databaseSequenceBatchSize = 20; /** * JDBC fetchSize hint when using findList. Defaults to 0 leaving it up to the JDBC driver. */ private int jdbcFetchSizeFindList; /** * JDBC fetchSize hint when using findEach/findEachWhile. Defaults to 100. Note that this does * not apply to MySql as that gets special treatment (forward only etc). */ private int jdbcFetchSizeFindEach = 100; /** * Suffix appended to the base table to derive the view that contains the union * of the base table and the history table in order to support asOf queries. */ private String asOfViewSuffix = "_with_history"; /** * Column used to support history and 'As of' queries. This column is a timestamp range * or equivalent. */ private String asOfSysPeriod = "sys_period"; /** * Suffix appended to the base table to derive the view that contains the union * of the base table and the history table in order to support asOf queries. */ private String historyTableSuffix = "_history"; /** * Use for transaction scoped batch mode. */ private PersistBatch persistBatch = PersistBatch.NONE; /** * Use for per request batch mode. */ private PersistBatch persistBatchOnCascade = PersistBatch.NONE; private int persistBatchSize = 20; /** * The default batch size for lazy loading */ private int lazyLoadBatchSize = 10; /** * The default batch size for 'query joins'. */ private int queryBatchSize = 100; private boolean eagerFetchLobs; /** * Timezone used to get/set Timestamp values via JDBC. */ private String dataTimeZone; private boolean ddlGenerate; private boolean ddlRun; private boolean ddlCreateOnly; private String ddlInitSql; private String ddlSeedSql; /** * When true L2 bean cache use is skipped after a write has occurred on a transaction. */ private boolean skipCacheAfterWrite = true; private boolean useJtaTransactionManager; /** * The external transaction manager (like Spring). */ private ExternalTransactionManager externalTransactionManager; /** * The data source (if programmatically provided). */ private DataSource dataSource; /** * The data source config. */ private DataSourceConfig dataSourceConfig = new DataSourceConfig(); /** * The db migration config (migration resource path etc). */ private DbMigrationConfig migrationConfig = new DbMigrationConfig(); /** * The ClassLoadConfig used to detect Joda, Java8, Jackson etc and create plugin instances given a className. */ private ClassLoadConfig classLoadConfig = new ClassLoadConfig(); /** * Set to true if the DataSource uses autoCommit. *

* Indicates that Ebean should use autoCommit friendly Transactions and TransactionManager. */ private boolean autoCommitMode; /** * Set to true if transaction begin should be started with explicit statement. */ private boolean explicitTransactionBeginMode; /** * The data source JNDI name if using a JNDI DataSource. */ private String dataSourceJndiName; /** * The database boolean true value (typically either 1, T, or Y). */ private String databaseBooleanTrue; /** * The database boolean false value (typically either 0, F or N). */ private String databaseBooleanFalse; /** * The naming convention. */ private NamingConvention namingConvention = new UnderscoreNamingConvention(); /** * Naming convention used in DDL generation for primary keys, foreign keys etc. */ private DbConstraintNaming constraintNaming = new DbConstraintNaming(); /** * Behaviour of update to include on the change properties. */ private boolean updateChangesOnly = true; /** * Behaviour of updates in JDBC batch to by default include all properties. */ private boolean updateAllPropertiesInBatch = true; /** * Default behaviour for updates when cascade save on a O2M or M2M to delete any missing children. */ private boolean updatesDeleteMissingChildren = true; /** * Setting to indicate if UUID should be stored as binary(16) or varchar(40) or native DB type (for H2 and Postgres). */ private DbUuid dbUuid = DbUuid.AUTO; private List idGenerators = new ArrayList(); private List findControllers = new ArrayList(); private List persistControllers = new ArrayList(); private List postLoaders = new ArrayList(); private List persistListeners = new ArrayList(); private List queryAdapters = new ArrayList(); private List bulkTableEventListeners = new ArrayList(); private List configStartupListeners = new ArrayList(); private List transactionEventListeners = new ArrayList(); /** * By default inserts are included in the change log. */ private boolean changeLogIncludeInserts = true; private ChangeLogPrepare changeLogPrepare; private ChangeLogListener changeLogListener; private ChangeLogRegister changeLogRegister; private ReadAuditLogger readAuditLogger; private ReadAuditPrepare readAuditPrepare; private EncryptKeyManager encryptKeyManager; private EncryptDeployManager encryptDeployManager; private Encryptor encryptor; private DbEncrypt dbEncrypt; private ServerCachePlugin serverCachePlugin; private ServerCacheManager serverCacheManager; private boolean collectQueryStatsByNode = true; private boolean collectQueryOrigins = true; /** * The default PersistenceContextScope used if one is not explicitly set on a query. */ private PersistenceContextScope persistenceContextScope = PersistenceContextScope.TRANSACTION; private JsonFactory jsonFactory; private boolean localTimeWithNanos; private boolean durationWithNanos; private int maxCallStack = 5; private boolean transactionRollbackOnChecked = true; private boolean registerJmxMBeans = true; // configuration for the background executor service (thread pool) private int backgroundExecutorSchedulePoolSize = 1; private int backgroundExecutorCorePoolSize = 1; private int backgroundExecutorMaxPoolSize = 8; private int backgroundExecutorIdleSecs = 60; private int backgroundExecutorShutdownSecs = 30; // defaults for the L2 bean caching private int cacheWarmingDelay = 30; private int cacheMaxSize = 10000; private int cacheMaxIdleTime = 600; private int cacheMaxTimeToLive = 60 * 60 * 6; // defaults for the L2 query caching private int queryCacheMaxSize = 1000; private int queryCacheMaxIdleTime = 600; private int queryCacheMaxTimeToLive = 60 * 60 * 6; private Object objectMapper; /** * Set to true if you want eq("someProperty", null) to generate 1=1 rather than "is null" sql expression. */ private boolean expressionEqualsWithNullAsNoop; /** * Set to true to use native ILIKE expression (if support by datasbase platform / like Postgres). */ private boolean expressionNativeIlike; private String jodaLocalTimeMode; /** * Set to true to globally disable L2 caching (typically for performance testing). */ private boolean disableL2Cache; /** * Construct a Server Configuration for programmatically creating an EbeanServer. */ public ServerConfig() { } /** * Put a service object into configuration such that it can be passed to a plugin. *

* For example, put IgniteConfiguration in to be passed to the Ignite plugin. *

*/ public void putServiceObject(String key, Object configObject) { serviceObject.put(key, configObject); } /** * Return the service object given the key. */ public Object getServiceObject(String key) { return serviceObject.get(key); } /** * Return the Jackson JsonFactory to use. *

* If not set a default implementation will be used. */ public JsonFactory getJsonFactory() { return jsonFactory; } /** * Set the Jackson JsonFactory to use. *

* If not set a default implementation will be used. */ public void setJsonFactory(JsonFactory jsonFactory) { this.jsonFactory = jsonFactory; } /** * Return the JSON format used for DateTime types. */ public JsonConfig.DateTime getJsonDateTime() { return jsonDateTime; } /** * Set the JSON format to use for DateTime types. */ public void setJsonDateTime(JsonConfig.DateTime jsonDateTime) { this.jsonDateTime = jsonDateTime; } /** * Return the JSON include mode used when writing JSON. */ public JsonConfig.Include getJsonInclude() { return jsonInclude; } /** * Set the JSON include mode used when writing JSON. *

* Set to NON_NULL or NON_EMPTY to suppress nulls or null & empty collections respectively. *

*/ public void setJsonInclude(JsonConfig.Include jsonInclude) { this.jsonInclude = jsonInclude; } /** * Return the name of the EbeanServer. */ public String getName() { return name; } /** * Set the name of the EbeanServer. */ public void setName(String name) { this.name = name; } /** * Return the container / clustering configuration. *

* The container holds all the EbeanServer instances and provides clustering communication * services to all the EbeanServer instances. */ public ContainerConfig getContainerConfig() { return containerConfig; } /** * Set the container / clustering configuration. *

* The container holds all the EbeanServer instances and provides clustering communication * services to all the EbeanServer instances. */ public void setContainerConfig(ContainerConfig containerConfig) { this.containerConfig = containerConfig; } /** * Return true if this server should be registered with the Ebean singleton * when it is created. *

* By default this is set to true. *

*/ public boolean isRegister() { return register; } /** * Set to false if you do not want this server to be registered with the Ebean * singleton when it is created. *

* By default this is set to true. *

*/ public void setRegister(boolean register) { this.register = register; } /** * Return true if this server should be registered as the "default" server * with the Ebean singleton. *

* This is only used when {@link #setRegister(boolean)} is also true. *

*/ public boolean isDefaultServer() { return defaultServer; } /** * Set true if this EbeanServer should be registered as the "default" server * with the Ebean singleton. *

* This is only used when {@link #setRegister(boolean)} is also true. *

*/ public void setDefaultServer(boolean defaultServer) { this.defaultServer = defaultServer; } /** * Return the CurrentUserProvider. This is used to populate @WhoCreated, @WhoModified and * support other audit features (who executed a query etc). */ public CurrentUserProvider getCurrentUserProvider() { return currentUserProvider; } /** * Set the CurrentUserProvider. This is used to populate @WhoCreated, @WhoModified and * support other audit features (who executed a query etc). */ public void setCurrentUserProvider(CurrentUserProvider currentUserProvider) { this.currentUserProvider = currentUserProvider; } /** * Return the PersistBatch mode to use by default at the transaction level. *

* When INSERT or ALL is used then save(), delete() etc do not execute immediately but instead go into * a JDBC batch execute buffer that is flushed. The buffer is flushed if a query is executed, transaction ends * or the batch size is meet. *

*/ public PersistBatch getPersistBatch() { return persistBatch; } /** * Set the JDBC batch mode to use at the transaction level. *

* When INSERT or ALL is used then save(), delete() etc do not execute immediately but instead go into * a JDBC batch execute buffer that is flushed. The buffer is flushed if a query is executed, transaction ends * or the batch size is meet. *

*/ public void setPersistBatch(PersistBatch persistBatch) { this.persistBatch = persistBatch; } /** * Return the JDBC batch mode to use per save(), delete(), insert() or update() request. *

* This makes sense when a save() or delete() cascades and executes multiple child statements. The best case * for this is when saving a master/parent bean this cascade inserts many detail/child beans. *

*

* This only takes effect when the persistBatch mode at the transaction level does not take effect. *

*/ public PersistBatch getPersistBatchOnCascade() { return persistBatchOnCascade; } /** * Set the JDBC batch mode to use per save(), delete(), insert() or update() request. *

* This makes sense when a save() or delete() etc cascades and executes multiple child statements. The best caase * for this is when saving a master/parent bean this cascade inserts many detail/child beans. *

*

* This only takes effect when the persistBatch mode at the transaction level does not take effect. *

*/ public void setPersistBatchOnCascade(PersistBatch persistBatchOnCascade) { this.persistBatchOnCascade = persistBatchOnCascade; } /** * Deprecated, please migrate to using setPersistBatch(). *

* Set to true if you what to use JDBC batching for persisting and deleting * beans. *

*

* With this Ebean will batch up persist requests and use the JDBC batch api. * This is a performance optimisation designed to reduce the network chatter. *

*

* When true this is equivalent to {@code setPersistBatch(PersistBatch.ALL)} or * when false to {@code setPersistBatch(PersistBatch.NONE)} *

*/ public void setPersistBatching(boolean persistBatching) { this.persistBatch = (persistBatching) ? PersistBatch.ALL : PersistBatch.NONE; } /** * Return the batch size used for JDBC batching. This defaults to 20. */ public int getPersistBatchSize() { return persistBatchSize; } /** * Set the batch size used for JDBC batching. If unset this defaults to 20. *

* You can also set the batch size on the transaction. *

* * @see com.avaje.ebean.Transaction#setBatchSize(int) */ public void setPersistBatchSize(int persistBatchSize) { this.persistBatchSize = persistBatchSize; } /** * Gets the query batch size. This defaults to 100. * * @return the query batch size */ public int getQueryBatchSize() { return queryBatchSize; } /** * Sets the query batch size. This defaults to 100. * * @param queryBatchSize the new query batch size */ public void setQueryBatchSize(int queryBatchSize) { this.queryBatchSize = queryBatchSize; } /** * Return the default batch size for lazy loading of beans and collections. */ public int getLazyLoadBatchSize() { return lazyLoadBatchSize; } /** * Set the default batch size for lazy loading. *

* This is the number of beans or collections loaded when lazy loading is * invoked by default. *

*

* The default value is for this is 10 (load 10 beans or collections). *

*

* You can explicitly control the lazy loading batch size for a given join on * a query using +lazy(batchSize) or JoinConfig. *

*/ public void setLazyLoadBatchSize(int lazyLoadBatchSize) { this.lazyLoadBatchSize = lazyLoadBatchSize; } /** * Set the number of sequences to fetch/preallocate when using DB sequences. *

* This is a performance optimisation to reduce the number times Ebean * requests a sequence to be used as an Id for a bean (aka reduce network * chatter). *

*/ public void setDatabaseSequenceBatchSize(int databaseSequenceBatchSize) { this.databaseSequenceBatchSize = databaseSequenceBatchSize; } /** * Return the default JDBC fetchSize hint for findList queries. */ public int getJdbcFetchSizeFindList() { return jdbcFetchSizeFindList; } /** * Set the default JDBC fetchSize hint for findList queries. */ public void setJdbcFetchSizeFindList(int jdbcFetchSizeFindList) { this.jdbcFetchSizeFindList = jdbcFetchSizeFindList; } /** * Return the default JDBC fetchSize hint for findEach/findEachWhile queries. */ public int getJdbcFetchSizeFindEach() { return jdbcFetchSizeFindEach; } /** * Set the default JDBC fetchSize hint for findEach/findEachWhile queries. */ public void setJdbcFetchSizeFindEach(int jdbcFetchSizeFindEach) { this.jdbcFetchSizeFindEach = jdbcFetchSizeFindEach; } /** * Return the ChangeLogPrepare. *

* This is used to set user context information to the ChangeSet in the * foreground thread prior to the logging occurring in a background thread. *

*/ public ChangeLogPrepare getChangeLogPrepare() { return changeLogPrepare; } /** * Set the ChangeLogPrepare. *

* This is used to set user context information to the ChangeSet in the * foreground thread prior to the logging occurring in a background thread. *

*/ public void setChangeLogPrepare(ChangeLogPrepare changeLogPrepare) { this.changeLogPrepare = changeLogPrepare; } /** * Return the ChangeLogListener which actually performs the logging of change sets * in the background. */ public ChangeLogListener getChangeLogListener() { return changeLogListener; } /** * Set the ChangeLogListener which actually performs the logging of change sets * in the background. */ public void setChangeLogListener(ChangeLogListener changeLogListener) { this.changeLogListener = changeLogListener; } /** * Return the ChangeLogRegister which controls which ChangeLogFilter is used for each * bean type and in this way provide fine grained control over which persist requests * are included in the change log. */ public ChangeLogRegister getChangeLogRegister() { return changeLogRegister; } /** * Set the ChangeLogRegister which controls which ChangeLogFilter is used for each * bean type and in this way provide fine grained control over which persist requests * are included in the change log. */ public void setChangeLogRegister(ChangeLogRegister changeLogRegister) { this.changeLogRegister = changeLogRegister; } /** * Return true if inserts should be included in the change log by default. */ public boolean isChangeLogIncludeInserts() { return changeLogIncludeInserts; } /** * Set if inserts should be included in the change log by default. */ public void setChangeLogIncludeInserts(boolean changeLogIncludeInserts) { this.changeLogIncludeInserts = changeLogIncludeInserts; } /** * Return the ReadAuditLogger to use. */ public ReadAuditLogger getReadAuditLogger() { return readAuditLogger; } /** * Set the ReadAuditLogger to use. If not set the default implementation is used * which logs the read events in JSON format to a standard named SLF4J logger * (which can be configured in say logback to log to a separate log file). */ public void setReadAuditLogger(ReadAuditLogger readAuditLogger) { this.readAuditLogger = readAuditLogger; } /** * Return the ReadAuditPrepare to use. */ public ReadAuditPrepare getReadAuditPrepare() { return readAuditPrepare; } /** * Set the ReadAuditPrepare to use. *

* It is expected that an implementation is used that read user context information * (user id, user ip address etc) and sets it on the ReadEvent bean before it is sent * to the ReadAuditLogger. *

*/ public void setReadAuditPrepare(ReadAuditPrepare readAuditPrepare) { this.readAuditPrepare = readAuditPrepare; } /** * Return the DB migration configuration. */ public DbMigrationConfig getMigrationConfig() { return migrationConfig; } /** * Set the DB migration configuration. */ public void setMigrationConfig(DbMigrationConfig migrationConfig) { this.migrationConfig = migrationConfig; } /** * Return the time zone to use when reading/writing Timestamps via JDBC. *

* When set a Calendar object is used in JDBC calls when reading/writing Timestamp objects. *

*/ public String getDataTimeZone() { return System.getProperty("ebean.dataTimeZone", dataTimeZone); } /** * Set the time zone to use when reading/writing Timestamps via JDBC. */ public void setDataTimeZone(String dataTimeZone) { this.dataTimeZone = dataTimeZone; } /** * Return the suffix appended to the base table to derive the view that contains the union * of the base table and the history table in order to support asOf queries. */ public String getAsOfViewSuffix() { return asOfViewSuffix; } /** * Set the suffix appended to the base table to derive the view that contains the union * of the base table and the history table in order to support asOf queries. */ public void setAsOfViewSuffix(String asOfViewSuffix) { this.asOfViewSuffix = asOfViewSuffix; } /** * Return the database column used to support history and 'As of' queries. This column is a timestamp range * or equivalent. */ public String getAsOfSysPeriod() { return asOfSysPeriod; } /** * Set the database column used to support history and 'As of' queries. This column is a timestamp range * or equivalent. */ public void setAsOfSysPeriod(String asOfSysPeriod) { this.asOfSysPeriod = asOfSysPeriod; } /** * Return the history table suffix (defaults to _history). */ public String getHistoryTableSuffix() { return historyTableSuffix; } /** * Set the history table suffix. */ public void setHistoryTableSuffix(String historyTableSuffix) { this.historyTableSuffix = historyTableSuffix; } /** * Return true if we are running in a JTA Transaction manager. */ public boolean isUseJtaTransactionManager() { return useJtaTransactionManager; } /** * Set to true if we are running in a JTA Transaction manager. */ public void setUseJtaTransactionManager(boolean useJtaTransactionManager) { this.useJtaTransactionManager = useJtaTransactionManager; } /** * Return the external transaction manager. */ public ExternalTransactionManager getExternalTransactionManager() { return externalTransactionManager; } /** * Set the external transaction manager. */ public void setExternalTransactionManager(ExternalTransactionManager externalTransactionManager) { this.externalTransactionManager = externalTransactionManager; } /** * Return the ServerCachePlugin. */ public ServerCachePlugin getServerCachePlugin() { return serverCachePlugin; } /** * Set the ServerCachePlugin to use. */ public void setServerCachePlugin(ServerCachePlugin serverCachePlugin) { this.serverCachePlugin = serverCachePlugin; } /** * Return the ServerCacheManager. */ public ServerCacheManager getServerCacheManager() { return serverCacheManager; } /** * Set the ServerCacheManager to use. */ public void setServerCacheManager(ServerCacheManager serverCacheManager) { this.serverCacheManager = serverCacheManager; } /** * Return true if LOB's should default to fetch eager. * By default this is set to false and LOB's must be explicitly fetched. */ public boolean isEagerFetchLobs() { return eagerFetchLobs; } /** * Set to true if you want LOB's to be fetch eager by default. * By default this is set to false and LOB's must be explicitly fetched. */ public void setEagerFetchLobs(boolean eagerFetchLobs) { this.eagerFetchLobs = eagerFetchLobs; } /** * Return the max call stack to use for origin location. */ public int getMaxCallStack() { return maxCallStack; } /** * Set the max call stack to use for origin location. */ public void setMaxCallStack(int maxCallStack) { this.maxCallStack = maxCallStack; } /** * Return true if transactions should rollback on checked exceptions. */ public boolean isTransactionRollbackOnChecked() { return transactionRollbackOnChecked; } /** * Set to true if transactions should by default rollback on checked exceptions. */ public void setTransactionRollbackOnChecked(boolean transactionRollbackOnChecked) { this.transactionRollbackOnChecked = transactionRollbackOnChecked; } /** * Return true if the server should register JMX MBeans. */ public boolean isRegisterJmxMBeans() { return registerJmxMBeans; } /** * Set if the server should register JMX MBeans. */ public void setRegisterJmxMBeans(boolean registerJmxMBeans) { this.registerJmxMBeans = registerJmxMBeans; } /** * Return the Background executor schedule pool size. Defaults to 1. */ public int getBackgroundExecutorSchedulePoolSize() { return backgroundExecutorSchedulePoolSize; } /** * Set the Background executor schedule pool size. */ public void setBackgroundExecutorSchedulePoolSize(int backgroundExecutorSchedulePoolSize) { this.backgroundExecutorSchedulePoolSize = backgroundExecutorSchedulePoolSize; } /** * Return the Background executor core pool size. */ public int getBackgroundExecutorCorePoolSize() { return backgroundExecutorCorePoolSize; } /** * Set the Background executor core pool size. */ public void setBackgroundExecutorCorePoolSize(int backgroundExecutorCorePoolSize) { this.backgroundExecutorCorePoolSize = backgroundExecutorCorePoolSize; } /** * Return the Background executor max pool size. */ public int getBackgroundExecutorMaxPoolSize() { return backgroundExecutorMaxPoolSize; } /** * Set the Background executor max pool size. */ public void setBackgroundExecutorMaxPoolSize(int backgroundExecutorMaxPoolSize) { this.backgroundExecutorMaxPoolSize = backgroundExecutorMaxPoolSize; } /** * Return the Background executor idle seconds. */ public int getBackgroundExecutorIdleSecs() { return backgroundExecutorIdleSecs; } /** * Set the Background executor idle seconds. */ public void setBackgroundExecutorIdleSecs(int backgroundExecutorIdleSecs) { this.backgroundExecutorIdleSecs = backgroundExecutorIdleSecs; } /** * Return the Background executor shutdown seconds. This is the time allowed for the pool to shutdown nicely * before it is forced shutdown. */ public int getBackgroundExecutorShutdownSecs() { return backgroundExecutorShutdownSecs; } /** * Set the Background executor shutdown seconds. This is the time allowed for the pool to shutdown nicely * before it is forced shutdown. */ public void setBackgroundExecutorShutdownSecs(int backgroundExecutorShutdownSecs) { this.backgroundExecutorShutdownSecs = backgroundExecutorShutdownSecs; } /** * Return the cache warming delay in seconds. */ public int getCacheWarmingDelay() { return cacheWarmingDelay; } /** * Set the cache warming delay in seconds. */ public void setCacheWarmingDelay(int cacheWarmingDelay) { this.cacheWarmingDelay = cacheWarmingDelay; } /** * Return the L2 cache default max size. */ public int getCacheMaxSize() { return cacheMaxSize; } /** * Set the L2 cache default max size. */ public void setCacheMaxSize(int cacheMaxSize) { this.cacheMaxSize = cacheMaxSize; } /** * Return the L2 cache default max idle time in seconds. */ public int getCacheMaxIdleTime() { return cacheMaxIdleTime; } /** * Set the L2 cache default max idle time in seconds. */ public void setCacheMaxIdleTime(int cacheMaxIdleTime) { this.cacheMaxIdleTime = cacheMaxIdleTime; } /** * Return the L2 cache default max time to live in seconds. */ public int getCacheMaxTimeToLive() { return cacheMaxTimeToLive; } /** * Set the L2 cache default max time to live in seconds. */ public void setCacheMaxTimeToLive(int cacheMaxTimeToLive) { this.cacheMaxTimeToLive = cacheMaxTimeToLive; } /** * Return the L2 query cache default max size. */ public int getQueryCacheMaxSize() { return queryCacheMaxSize; } /** * Set the L2 query cache default max size. */ public void setQueryCacheMaxSize(int queryCacheMaxSize) { this.queryCacheMaxSize = queryCacheMaxSize; } /** * Return the L2 query cache default max idle time in seconds. */ public int getQueryCacheMaxIdleTime() { return queryCacheMaxIdleTime; } /** * Set the L2 query cache default max idle time in seconds. */ public void setQueryCacheMaxIdleTime(int queryCacheMaxIdleTime) { this.queryCacheMaxIdleTime = queryCacheMaxIdleTime; } /** * Return the L2 query cache default max time to live in seconds. */ public int getQueryCacheMaxTimeToLive() { return queryCacheMaxTimeToLive; } /** * Set the L2 query cache default max time to live in seconds. */ public void setQueryCacheMaxTimeToLive(int queryCacheMaxTimeToLive) { this.queryCacheMaxTimeToLive = queryCacheMaxTimeToLive; } /** * Return the NamingConvention. *

* If none has been set the default UnderscoreNamingConvention is used. *

*/ public NamingConvention getNamingConvention() { return namingConvention; } /** * Set the NamingConvention. *

* If none is set the default UnderscoreNamingConvention is used. *

*/ public void setNamingConvention(NamingConvention namingConvention) { this.namingConvention = namingConvention; } /** * Return the configuration for the ElasticSearch integration. */ public DocStoreConfig getDocStoreConfig() { return docStoreConfig; } /** * Set the configuration for the ElasticSearch integration. */ public void setDocStoreConfig(DocStoreConfig docStoreConfig) { this.docStoreConfig = docStoreConfig; } /** * Return the constraint naming convention used in DDL generation. */ public DbConstraintNaming getConstraintNaming() { return constraintNaming; } /** * Set the constraint naming convention used in DDL generation. */ public void setConstraintNaming(DbConstraintNaming constraintNaming) { this.constraintNaming = constraintNaming; } /** * Return the configuration for AutoTune. */ public AutoTuneConfig getAutoTuneConfig() { return autoTuneConfig; } /** * Set the configuration for AutoTune. */ public void setAutoTuneConfig(AutoTuneConfig autoTuneConfig) { this.autoTuneConfig = autoTuneConfig; } /** * Return the DataSource. */ public DataSource getDataSource() { return dataSource; } /** * Set a DataSource. */ public void setDataSource(DataSource dataSource) { this.dataSource = dataSource; } /** * Return the configuration to build a DataSource using Ebean's own DataSource * implementation. */ public DataSourceConfig getDataSourceConfig() { return dataSourceConfig; } /** * Set the configuration required to build a DataSource using Ebean's own * DataSource implementation. */ public void setDataSourceConfig(DataSourceConfig dataSourceConfig) { this.dataSourceConfig = dataSourceConfig; } /** * Return the JNDI name of the DataSource to use. */ public String getDataSourceJndiName() { return dataSourceJndiName; } /** * Set the JNDI name of the DataSource to use. *

* By default a prefix of "java:comp/env/jdbc/" is used to lookup the * DataSource. This prefix is not used if dataSourceJndiName starts with * "java:". *

*/ public void setDataSourceJndiName(String dataSourceJndiName) { this.dataSourceJndiName = dataSourceJndiName; } /** * Return true if autoCommit mode is on. This indicates to Ebean to use autoCommit friendly Transactions and TransactionManager. */ public boolean isAutoCommitMode() { return autoCommitMode; } /** * Set to true if autoCommit mode is on and Ebean should use autoCommit friendly Transactions and TransactionManager. */ public void setAutoCommitMode(boolean autoCommitMode) { this.autoCommitMode = autoCommitMode; } /** * Return true if transaction begin should be started with explicit statement. */ public boolean isExplicitTransactionBeginMode() { return explicitTransactionBeginMode; } /** * Set to true if transaction begin should be started with explicit statement. *

* This works for H2 and Postgres but not for Oracle - only use this if you first name * is Daryl or you have explicitly talked to Rob about this feature. *

*

* This is generally not expected to be turned on but instead allow transactions to start * implicitly which is generally the standard approach. *

*/ public void setExplicitTransactionBeginMode(boolean explicitTransactionBeginMode) { this.explicitTransactionBeginMode = explicitTransactionBeginMode; } /** * Return a value used to represent TRUE in the database. *

* This is used for databases that do not support boolean natively. *

*

* The value returned is either a Integer or a String (e.g. "1", or "T"). *

*/ public String getDatabaseBooleanTrue() { return databaseBooleanTrue; } /** * Set the value to represent TRUE in the database. *

* This is used for databases that do not support boolean natively. *

*

* The value set is either a Integer or a String (e.g. "1", or "T"). *

*/ public void setDatabaseBooleanTrue(String databaseTrue) { this.databaseBooleanTrue = databaseTrue; } /** * Return a value used to represent FALSE in the database. *

* This is used for databases that do not support boolean natively. *

*

* The value returned is either a Integer or a String (e.g. "0", or "F"). *

*/ public String getDatabaseBooleanFalse() { return databaseBooleanFalse; } /** * Set the value to represent FALSE in the database. *

* This is used for databases that do not support boolean natively. *

*

* The value set is either a Integer or a String (e.g. "0", or "F"). *

*/ public void setDatabaseBooleanFalse(String databaseFalse) { this.databaseBooleanFalse = databaseFalse; } /** * Return the number of DB sequence values that should be preallocated. */ public int getDatabaseSequenceBatchSize() { return databaseSequenceBatchSize; } /** * Set the number of DB sequence values that should be preallocated and cached * by Ebean. *

* This is only used for DB's that use sequences and is a performance * optimisation. This reduces the number of times Ebean needs to get a * sequence value from the Database reducing network chatter. *

*

* By default this value is 10 so when we need another Id (and don't have one * in our cache) Ebean will fetch 10 id's from the database. Note that when * the cache drops to have full (which is 5 by default) Ebean will fetch * another batch of Id's in a background thread. *

*/ public void setDatabaseSequenceBatch(int databaseSequenceBatchSize) { this.databaseSequenceBatchSize = databaseSequenceBatchSize; } /** * Return the database platform name (can be null). *

* If null then the platform is determined automatically via the JDBC driver * information. *

*/ public String getDatabasePlatformName() { return databasePlatformName; } /** * Explicitly set the database platform name *

* If none is set then the platform is determined automatically via the JDBC * driver information. *

*

* This can be used when the Database Platform can not be automatically * detected from the JDBC driver (possibly 3rd party JDBC driver). It is also * useful when you want to do offline DDL generation for a database platform * that you don't have access to. *

*

* Values are oracle, h2, postgres, mysql, mssqlserver2005. *

*/ public void setDatabasePlatformName(String databasePlatformName) { this.databasePlatformName = databasePlatformName; } /** * Return the database platform to use for this server. */ public DatabasePlatform getDatabasePlatform() { return databasePlatform; } /** * Explicitly set the database platform to use. *

* If none is set then the platform is determined via the databasePlatformName * or automatically via the JDBC driver information. *

*/ public void setDatabasePlatform(DatabasePlatform databasePlatform) { this.databasePlatform = databasePlatform; } /** * Return the EncryptKeyManager. */ public EncryptKeyManager getEncryptKeyManager() { return encryptKeyManager; } /** * Set the EncryptKeyManager. *

* This is required when you want to use encrypted properties. *

*

* You can also set this in ebean.proprerties: *

* *
{@code
   * # set via ebean.properties
   * ebean.encryptKeyManager=com.avaje.tests.basic.encrypt.BasicEncyptKeyManager
   * }
*/ public void setEncryptKeyManager(EncryptKeyManager encryptKeyManager) { this.encryptKeyManager = encryptKeyManager; } /** * Return the EncryptDeployManager. *

* This is optionally used to programmatically define which columns are * encrypted instead of using the {@link Encrypted} Annotation. *

*/ public EncryptDeployManager getEncryptDeployManager() { return encryptDeployManager; } /** * Set the EncryptDeployManager. *

* This is optionally used to programmatically define which columns are * encrypted instead of using the {@link Encrypted} Annotation. *

*/ public void setEncryptDeployManager(EncryptDeployManager encryptDeployManager) { this.encryptDeployManager = encryptDeployManager; } /** * Return the Encryptor used to encrypt data on the java client side (as * opposed to DB encryption functions). */ public Encryptor getEncryptor() { return encryptor; } /** * Set the Encryptor used to encrypt data on the java client side (as opposed * to DB encryption functions). *

* Ebean has a default implementation that it will use if you do not set your * own Encryptor implementation. *

*/ public void setEncryptor(Encryptor encryptor) { this.encryptor = encryptor; } /** * Return the DbEncrypt used to encrypt and decrypt properties. *

* Note that if this is not set then the DbPlatform may already have a * DbEncrypt set and that will be used. *

*/ public DbEncrypt getDbEncrypt() { return dbEncrypt; } /** * Set the DbEncrypt used to encrypt and decrypt properties. *

* Note that if this is not set then the DbPlatform may already have a * DbEncrypt set (H2, MySql, Postgres and Oracle platforms have a DbEncrypt) *

*/ public void setDbEncrypt(DbEncrypt dbEncrypt) { this.dbEncrypt = dbEncrypt; } /** * Return the DB type used to store UUID. */ public DbUuid getDbUuid() { return dbUuid; } /** * Set the DB type used to store UUID. */ public void setDbUuid(DbUuid dbUuid) { this.dbUuid = dbUuid; } /** * Return true if LocalTime should be persisted with nanos precision. */ public boolean isLocalTimeWithNanos() { return localTimeWithNanos; } /** * Set to true if LocalTime should be persisted with nanos precision. *

* Otherwise it is persisted using java.sql.Time which is seconds precision. *

*/ public void setLocalTimeWithNanos(boolean localTimeWithNanos) { this.localTimeWithNanos = localTimeWithNanos; } /** * Return true if Duration should be persisted with nanos precision (SQL DECIMAL). *

* Otherwise it is persisted with second precision (SQL INTEGER). *

*/ public boolean isDurationWithNanos() { return durationWithNanos; } /** * Set to true if Duration should be persisted with nanos precision (SQL DECIMAL). *

* Otherwise it is persisted with second precision (SQL INTEGER). *

*/ public void setDurationWithNanos(boolean durationWithNanos) { this.durationWithNanos = durationWithNanos; } /** * Set to true to generate the "create all" DDL on startup. *

* Typically we want this on when we are running tests locally (and often using H2) * and we want to create the full DB schema from scratch to run tests. */ public void setDdlGenerate(boolean ddlGenerate) { this.ddlGenerate = ddlGenerate; } /** * Set to true to run the generated "create all DDL" on startup. *

* Typically we want this on when we are running tests locally (and often using H2) * and we want to create the full DB schema from scratch to run tests. */ public void setDdlRun(boolean ddlRun) { this.ddlRun = ddlRun; } /** * Return true if the "drop all ddl" should be skipped. *

* Typically we want to do this when using H2 (in memory) as our test database and the drop statements * are not required so skipping the drop table statements etc makes it faster with less noise in the logs. */ public boolean isDdlCreateOnly() { return ddlCreateOnly; } /** * Set to true if the "drop all ddl" should be skipped. *

* Typically we want to do this when using H2 (in memory) as our test database and the drop statements * are not required so skipping the drop table statements etc makes it faster with less noise in the logs. */ public void setDdlCreateOnly(boolean ddlCreateOnly) { this.ddlCreateOnly = ddlCreateOnly; } /** * Return SQL script to execute after the "create all" DDL has been run. *

* Typically this is a sql script that inserts test seed data when running tests. * Place a sql script in src/test/resources that inserts test seed data. *

*/ public String getDdlSeedSql() { return ddlSeedSql; } /** * Set a SQL script to execute after the "create all" DDL has been run. *

* Typically this is a sql script that inserts test seed data when running tests. * Place a sql script in src/test/resources that inserts test seed data. *

*/ public void setDdlSeedSql(String ddlSeedSql) { this.ddlSeedSql = ddlSeedSql; } /** * Return a SQL script to execute before the "create all" DDL has been run. */ public String getDdlInitSql() { return ddlInitSql; } /** * Set a SQL script to execute before the "create all" DDL has been run. */ public void setDdlInitSql(String ddlInitSql) { this.ddlInitSql = ddlInitSql; } /** * Return true if the DDL should be generated. */ public boolean isDdlGenerate() { return ddlGenerate; } /** * Return true if the DDL should be run. */ public boolean isDdlRun() { return ddlRun; } /** * Return true if the class path search should be disabled. */ public boolean isDisableClasspathSearch() { return disableClasspathSearch; } /** * Set to true to disable the class path search even for the case where no entity bean classes * have been registered. This can be used to start an EbeanServer instance just to use the * SQL functions such as SqlQuery, SqlUpdate etc. */ public void setDisableClasspathSearch(boolean disableClasspathSearch) { this.disableClasspathSearch = disableClasspathSearch; } /** * Return the mode to use for Joda LocalTime support 'normal' or 'utc'. */ public String getJodaLocalTimeMode() { return jodaLocalTimeMode; } /** * Set the mode to use for Joda LocalTime support 'normal' or 'utc'. */ public void setJodaLocalTimeMode(String jodaLocalTimeMode) { this.jodaLocalTimeMode = jodaLocalTimeMode; } /** * Programmatically add classes (typically entities) that this server should * use. *

* The class can be an Entity, Embedded type, ScalarType, BeanPersistListener, * BeanFinder or BeanPersistController. *

*

* If no classes are specified then the classes are found automatically via * searching the class path. *

*

* Alternatively the classes can be added via {@link #setClasses(List)}. *

* * @param cls the entity type (or other type) that should be registered by this * server. */ public void addClass(Class cls) { if (classes == null) { classes = new ArrayList>(); } classes.add(cls); } /** * Add a package to search for entities via class path search. *

* This is only used if classes have not been explicitly specified. *

*/ public void addPackage(String packageName) { if (packages == null) { packages = new ArrayList(); } packages.add(packageName); } /** * Return packages to search for entities via class path search. *

* This is only used if classes have not been explicitly specified. *

*/ public List getPackages() { return packages; } /** * Set packages to search for entities via class path search. *

* This is only used if classes have not been explicitly specified. *

*/ public void setPackages(List packages) { this.packages = packages; } /** * Set the list of classes (entities, listeners, scalarTypes etc) that should * be used for this server. *

* If no classes are specified then the classes are found automatically via * searching the class path. *

*

* Alternatively the classes can contain added via {@link #addClass(Class)}. *

*/ public void setClasses(List> classes) { this.classes = classes; } /** * Return the classes registered for this server. Typically this includes * entities and perhaps listeners. */ public List> getClasses() { return classes; } /** * Return true if L2 bean cache should be skipped once writes have occurred on a transaction. *

* This defaults to true and means that for "find by id" and "find by natural key" * queries that normally hit L2 bean cache automatically will not do so after a write/persist * on the transaction. *

* *
{@code
   *
   *   // assume Customer has L2 bean caching enabled ...
   *
   *   Transaction transaction = Ebean.beginTransaction();
   *   try {
   *
   *     // this uses L2 bean cache as the transaction
   *     // ... is considered "query only" at this point
   *     Customer.find.byId(42);
   *
   *     // transaction no longer "query only" once
   *     // ... a bean has been saved etc
   *     Ebean.save(someBean);
   *
   *     // will NOT use L2 bean cache as the transaction
   *     // ... is no longer considered "query only"
   *     Customer.find.byId(55);
   *
   *
   *
   *     // explicit control - please use L2 bean cache
   *
   *     transaction.setSkipCache(false);
   *     Customer.find.byId(77); // hit the l2 bean cache
   *
   *
   *     // explicit control - please don't use L2 bean cache
   *
   *     transaction.setSkipCache(true);
   *     Customer.find.byId(99); // skips l2 bean cache
   *
   *
   *   } finally {
   *     transaction.end();
   *   }
   *
   * }
* * @see com.avaje.ebean.Transaction#setSkipCache(boolean) */ public boolean isSkipCacheAfterWrite() { return skipCacheAfterWrite; } /** * Set to false when we still want to hit the cache after a write has occurred on a transaction. */ public void setSkipCacheAfterWrite(boolean skipCacheAfterWrite) { this.skipCacheAfterWrite = skipCacheAfterWrite; } /** * Return true to only update changed properties. */ public boolean isUpdateChangesOnly() { return updateChangesOnly; } /** * Set to true to only update changed properties. */ public void setUpdateChangesOnly(boolean updateChangesOnly) { this.updateChangesOnly = updateChangesOnly; } /** * Returns true if updates in JDBC batch default to include all properties by default. */ public boolean isUpdateAllPropertiesInBatch() { return updateAllPropertiesInBatch; } /** * Set to false if by default updates in JDBC batch should not include all properties. *

* This mode can be explicitly set per transaction. *

* * @see com.avaje.ebean.Transaction#setUpdateAllLoadedProperties(boolean) */ public void setUpdateAllPropertiesInBatch(boolean updateAllPropertiesInBatch) { this.updateAllPropertiesInBatch = updateAllPropertiesInBatch; } /** * Return true if updates by default delete missing children when cascading save to a OneToMany or * ManyToMany. When not set this defaults to true. */ public boolean isUpdatesDeleteMissingChildren() { return updatesDeleteMissingChildren; } /** * Set if updates by default delete missing children when cascading save to a OneToMany or * ManyToMany. When not set this defaults to true. */ public void setUpdatesDeleteMissingChildren(boolean updatesDeleteMissingChildren) { this.updatesDeleteMissingChildren = updatesDeleteMissingChildren; } /** * Return true if the ebeanServer should collection query statistics by ObjectGraphNode. */ public boolean isCollectQueryStatsByNode() { return collectQueryStatsByNode; } /** * Set to true to collection query execution statistics by ObjectGraphNode. *

* These statistics can be used to highlight code/query 'origin points' that result in lots of lazy loading. *

*

* It is considered safe/fine to have this set to true for production. *

*

* This information can be later retrieved via {@link MetaInfoManager}. *

* * @see MetaInfoManager */ public void setCollectQueryStatsByNode(boolean collectQueryStatsByNode) { this.collectQueryStatsByNode = collectQueryStatsByNode; } /** * Return true if query plans should also collect their 'origins'. This means for a given query plan you * can identify the code/origin points where this query resulted from including lazy loading origins. */ public boolean isCollectQueryOrigins() { return collectQueryOrigins; } /** * Set to true if query plans should collect their 'origin' points. This means for a given query plan you * can identify the code/origin points where this query resulted from including lazy loading origins. *

* This information can be later retrieved via {@link MetaInfoManager}. *

* * @see MetaInfoManager */ public void setCollectQueryOrigins(boolean collectQueryOrigins) { this.collectQueryOrigins = collectQueryOrigins; } /** * Returns the resource directory. */ public String getResourceDirectory() { return resourceDirectory; } /** * Sets the resource directory. */ public void setResourceDirectory(String resourceDirectory) { this.resourceDirectory = resourceDirectory; } /** * Register a BeanQueryAdapter instance. *

* Note alternatively you can use {@link #setQueryAdapters(List)} to set all * the BeanQueryAdapter instances. *

*/ public void add(BeanQueryAdapter beanQueryAdapter) { queryAdapters.add(beanQueryAdapter); } /** * Return the BeanQueryAdapter instances. */ public List getQueryAdapters() { return queryAdapters; } /** * Register all the BeanQueryAdapter instances. *

* Note alternatively you can use {@link #add(BeanQueryAdapter)} to add * BeanQueryAdapter instances one at a time. *

*/ public void setQueryAdapters(List queryAdapters) { this.queryAdapters = queryAdapters; } /** * Return the custom IdGenerator instances. */ public List getIdGenerators() { return idGenerators; } /** * Set the custom IdGenerator instances. */ public void setIdGenerators(List idGenerators) { this.idGenerators = idGenerators; } /** * Register a customer IdGenerator instance. */ public void add(IdGenerator idGenerator) { idGenerators.add(idGenerator); } /** * Register a BeanPersistController instance. *

* Note alternatively you can use {@link #setPersistControllers(List)} to set * all the BeanPersistController instances. *

*/ public void add(BeanPersistController beanPersistController) { persistControllers.add(beanPersistController); } /** * Register a BeanPostLoad instance. *

* Note alternatively you can use {@link #setPostLoaders(List)} to set * all the BeanPostLoad instances. *

*/ public void add(BeanPostLoad postLoad) { postLoaders.add(postLoad); } /** * Return the list of BeanFindController instances. */ public List getFindControllers() { return findControllers; } /** * Set the list of BeanFindController instances. */ public void setFindControllers(List findControllers) { this.findControllers = findControllers; } /** * Return the list of BeanPostLoader instances. */ public List getPostLoaders() { return postLoaders; } /** * Set the list of BeanPostLoader instances. */ public void setPostLoaders(List postLoaders) { this.postLoaders = postLoaders; } /** * Return the BeanPersistController instances. */ public List getPersistControllers() { return persistControllers; } /** * Register all the BeanPersistController instances. *

* Note alternatively you can use {@link #add(BeanPersistController)} to add * BeanPersistController instances one at a time. *

*/ public void setPersistControllers(List persistControllers) { this.persistControllers = persistControllers; } /** * Register a TransactionEventListener instance *

* Note alternatively you can use {@link #setTransactionEventListeners(List)} * to set all the TransactionEventListener instances. *

*/ public void add(TransactionEventListener listener) { transactionEventListeners.add(listener); } /** * Return the TransactionEventListener instances. */ public List getTransactionEventListeners() { return transactionEventListeners; } /** * Register all the TransactionEventListener instances. *

* Note alternatively you can use {@link #add(TransactionEventListener)} to * add TransactionEventListener instances one at a time. *

*/ public void setTransactionEventListeners(List transactionEventListeners) { this.transactionEventListeners = transactionEventListeners; } /** * Register a BeanPersistListener instance. *

* Note alternatively you can use {@link #setPersistListeners(List)} to set * all the BeanPersistListener instances. *

*/ public void add(BeanPersistListener beanPersistListener) { persistListeners.add(beanPersistListener); } /** * Return the BeanPersistListener instances. */ public List getPersistListeners() { return persistListeners; } /** * Add a BulkTableEventListener */ public void add(BulkTableEventListener bulkTableEventListener) { bulkTableEventListeners.add(bulkTableEventListener); } /** * Return the list of BulkTableEventListener instances. */ public List getBulkTableEventListeners() { return bulkTableEventListeners; } /** * Add a ServerConfigStartup. */ public void addServerConfigStartup(ServerConfigStartup configStartupListener) { configStartupListeners.add(configStartupListener); } /** * Return the list of ServerConfigStartup instances. */ public List getServerConfigStartupListeners() { return configStartupListeners; } /** * Register all the BeanPersistListener instances. *

* Note alternatively you can use {@link #add(BeanPersistListener)} to add * BeanPersistListener instances one at a time. *

*/ public void setPersistListeners(List persistListeners) { this.persistListeners = persistListeners; } /** * Return the default PersistenceContextScope to be used if one is not explicitly set on a query. *

* The PersistenceContextScope can specified on each query via {@link com.avaje.ebean * .Query#setPersistenceContextScope(com.avaje.ebean.PersistenceContextScope)}. If it * is not set on the query this default scope is used. * * @see com.avaje.ebean.Query#setPersistenceContextScope(com.avaje.ebean.PersistenceContextScope) */ public PersistenceContextScope getPersistenceContextScope() { // if somehow null return TRANSACTION scope return persistenceContextScope == null ? PersistenceContextScope.TRANSACTION : persistenceContextScope; } /** * Set the PersistenceContext scope to be used if one is not explicitly set on a query. *

* This defaults to {@link PersistenceContextScope#TRANSACTION}. *

* The PersistenceContextScope can specified on each query via {@link com.avaje.ebean * .Query#setPersistenceContextScope(com.avaje.ebean.PersistenceContextScope)}. If it * is not set on the query this scope is used. * * @see com.avaje.ebean.Query#setPersistenceContextScope(com.avaje.ebean.PersistenceContextScope) */ public void setPersistenceContextScope(PersistenceContextScope persistenceContextScope) { this.persistenceContextScope = persistenceContextScope; } /** * Return the ClassLoadConfig which is used to detect Joda, Java8 types etc and also * create new instances of plugins given a className. */ public ClassLoadConfig getClassLoadConfig() { return classLoadConfig; } /** * Set the ClassLoadConfig which is used to detect Joda, Java8 types etc and also * create new instances of plugins given a className. */ public void setClassLoadConfig(ClassLoadConfig classLoadConfig) { this.classLoadConfig = classLoadConfig; } /** * Return the service loader using the classLoader defined in ClassLoadConfig. */ public ServiceLoader serviceLoad(Class spiService) { return ServiceLoader.load(spiService, classLoadConfig.getClassLoader()); } /** * Return the first service using the service loader (or null). */ public T service(Class spiService) { ServiceLoader load = serviceLoad(spiService); Iterator serviceInstances = load.iterator(); return serviceInstances.hasNext() ? serviceInstances.next() : null; } /** * Load settings from ebean.properties. */ public void loadFromProperties() { loadFromProperties(PropertyMap.defaultProperties()); } /** * Load the settings from the given properties */ public void loadFromProperties(Properties properties) { // keep the properties used for configuration so that these are available for plugins this.properties = properties; PropertiesWrapper p = new PropertiesWrapper("ebean", name, properties); loadSettings(p); } /** * Load settings from test-ebean.properties and do nothing if the properties is not found. *

* This is typically used when test-ebean.properties is put into the test class path and used * to configure Ebean for running tests. *

*/ public void loadTestProperties() { Properties properties = PropertyMap.testProperties(); if (!properties.isEmpty()) { PropertiesWrapper p = new PropertiesWrapper("ebean", name, properties); loadSettings(p); } } /** * Return the properties that we used for configuration and were set via a call to loadFromProperties(). */ public Properties getProperties() { return properties; } /** * Return the instance to use (can be null) for the given plugin. * * @param properties the properties * @param pluginType the type of plugin * @param key properties key * @param instance existing instance */ @SuppressWarnings("unchecked") protected T createInstance(PropertiesWrapper properties, Class pluginType, String key, T instance) { if (instance != null) { return instance; } String classname = properties.get(key, null); return createInstance(pluginType, classname); } /** * Return the instance to use (can be null) for the given plugin. * * @param pluginType the type of plugin * @param classname the implementation class as per properties */ protected T createInstance(Class pluginType, String classname) { return classname == null ? null : (T) classLoadConfig.newInstance(classname); } /** * loads the data source settings to preserve existing behaviour. IMHO, if someone has set the datasource config already, * they don't want the settings to be reloaded and reset. This allows a descending class to override this behaviour and prevent it * from happening. * * @param p - The defined property source passed to load settings */ protected void loadDataSourceSettings(PropertiesWrapper p) { dataSourceConfig.loadSettings(p.properties, name); } /** * This is broken out to allow overridden behaviour. */ protected void loadDocStoreSettings(PropertiesWrapper p) { docStoreConfig.loadSettings(p); } /** * This is broken out to allow overridden behaviour. */ protected void loadAutoTuneSettings(PropertiesWrapper p) { autoTuneConfig.loadSettings(p); } /** * Load the configuration settings from the properties file. */ protected void loadSettings(PropertiesWrapper p) { migrationConfig.loadSettings(p, name); namingConvention = createNamingConvention(p, namingConvention); if (namingConvention != null) { namingConvention.loadFromProperties(p); } if (autoTuneConfig == null) { autoTuneConfig = new AutoTuneConfig(); } loadAutoTuneSettings(p); if (dataSourceConfig == null) { dataSourceConfig = new DataSourceConfig(); } loadDataSourceSettings(p); if (docStoreConfig == null) { docStoreConfig = new DocStoreConfig(); } loadDocStoreSettings(p); disableL2Cache = p.getBoolean("disableL2Cache", disableL2Cache); explicitTransactionBeginMode = p.getBoolean("explicitTransactionBeginMode", explicitTransactionBeginMode); autoCommitMode = p.getBoolean("autoCommitMode", autoCommitMode); useJtaTransactionManager = p.getBoolean("useJtaTransactionManager", useJtaTransactionManager); disableClasspathSearch = p.getBoolean("disableClasspathSearch", disableClasspathSearch); currentUserProvider = createInstance(p, CurrentUserProvider.class, "currentUserProvider", currentUserProvider); databasePlatform = createInstance(p, DatabasePlatform.class, "databasePlatform", databasePlatform); encryptKeyManager = createInstance(p, EncryptKeyManager.class, "encryptKeyManager", encryptKeyManager); encryptDeployManager = createInstance(p, EncryptDeployManager.class, "encryptDeployManager", encryptDeployManager); encryptor = createInstance(p, Encryptor.class, "encryptor", encryptor); dbEncrypt = createInstance(p, DbEncrypt.class, "dbEncrypt", dbEncrypt); serverCachePlugin = createInstance(p, ServerCachePlugin.class, "serverCachePlugin", serverCachePlugin); serverCacheManager = createInstance(p, ServerCacheManager.class, "serverCacheManager", serverCacheManager); cacheWarmingDelay = p.getInt("cacheWarmingDelay", cacheWarmingDelay); if (packages != null) { String packagesProp = p.get("search.packages", p.get("packages", null)); packages = getSearchJarsPackages(packagesProp); } collectQueryStatsByNode = p.getBoolean("collectQueryStatsByNode", collectQueryStatsByNode); collectQueryOrigins = p.getBoolean("collectQueryOrigins", collectQueryOrigins); skipCacheAfterWrite = p.getBoolean("skipCacheAfterWrite", skipCacheAfterWrite); updateAllPropertiesInBatch = p.getBoolean("updateAllPropertiesInBatch", updateAllPropertiesInBatch); updateChangesOnly = p.getBoolean("updateChangesOnly", updateChangesOnly); boolean defaultDeleteMissingChildren = p.getBoolean("defaultDeleteMissingChildren", updatesDeleteMissingChildren); updatesDeleteMissingChildren = p.getBoolean("updatesDeleteMissingChildren", defaultDeleteMissingChildren); if (p.get("batch.mode") != null || p.get("persistBatching") != null) { throw new IllegalArgumentException("Property 'batch.mode' or 'persistBatching' is being set but no longer used. Please change to use 'persistBatchMode'"); } persistBatch = p.getEnum(PersistBatch.class, "persistBatch", persistBatch); persistBatchOnCascade = p.getEnum(PersistBatch.class, "persistBatchOnCascade", persistBatchOnCascade); int batchSize = p.getInt("batch.size", persistBatchSize); persistBatchSize = p.getInt("persistBatchSize", batchSize); persistenceContextScope = PersistenceContextScope.valueOf(p.get("persistenceContextScope", "TRANSACTION")); changeLogIncludeInserts = p.getBoolean("changeLogIncludeInserts", changeLogIncludeInserts); expressionEqualsWithNullAsNoop = p.getBoolean("expressionEqualsWithNullAsNoop", expressionEqualsWithNullAsNoop); expressionNativeIlike = p.getBoolean("expressionNativeIlike", expressionNativeIlike); dataTimeZone = p.get("dataTimeZone", dataTimeZone); asOfViewSuffix = p.get("asOfViewSuffix", asOfViewSuffix); asOfSysPeriod = p.get("asOfSysPeriod", asOfSysPeriod); historyTableSuffix = p.get("historyTableSuffix", historyTableSuffix); dataSourceJndiName = p.get("dataSourceJndiName", dataSourceJndiName); jdbcFetchSizeFindEach = p.getInt("jdbcFetchSizeFindEach", jdbcFetchSizeFindEach); jdbcFetchSizeFindList = p.getInt("jdbcFetchSizeFindList", jdbcFetchSizeFindList); databaseSequenceBatchSize = p.getInt("databaseSequenceBatchSize", databaseSequenceBatchSize); databaseBooleanTrue = p.get("databaseBooleanTrue", databaseBooleanTrue); databaseBooleanFalse = p.get("databaseBooleanFalse", databaseBooleanFalse); databasePlatformName = p.get("databasePlatformName", databasePlatformName); dbUuid = p.getEnum(DbUuid.class, "dbuuid", dbUuid); if (p.getBoolean("uuidStoreAsBinary", false)) { dbUuid = DbUuid.BINARY; } localTimeWithNanos = p.getBoolean("localTimeWithNanos", localTimeWithNanos); jodaLocalTimeMode = p.get("jodaLocalTimeMode", jodaLocalTimeMode); lazyLoadBatchSize = p.getInt("lazyLoadBatchSize", lazyLoadBatchSize); queryBatchSize = p.getInt("queryBatchSize", queryBatchSize); jsonInclude = p.getEnum(JsonConfig.Include.class, "jsonInclude", jsonInclude); String jsonDateTimeFormat = p.get("jsonDateTime", null); if (jsonDateTimeFormat != null) { jsonDateTime = JsonConfig.DateTime.valueOf(jsonDateTimeFormat); } else { jsonDateTime = JsonConfig.DateTime.MILLIS; } ddlGenerate = p.getBoolean("ddl.generate", ddlGenerate); ddlRun = p.getBoolean("ddl.run", ddlRun); ddlCreateOnly = p.getBoolean("ddl.createOnly", ddlCreateOnly); ddlInitSql = p.get("ddl.initSql", ddlInitSql); ddlSeedSql = p.get("ddl.seedSql", ddlSeedSql); classes = getClasses(p); } private NamingConvention createNamingConvention(PropertiesWrapper properties, NamingConvention namingConvention) { NamingConvention nc = createInstance(properties, NamingConvention.class, "namingconvention", null); return (nc != null) ? nc : namingConvention; } /** * Build the list of classes from the comma delimited string. * * @param properties the properties * @return the classes */ private List> getClasses(PropertiesWrapper properties) { String classNames = properties.get("classes", null); if (classNames == null) { return classes; } List> classes = new ArrayList>(); String[] split = classNames.split("[ ,;]"); for (int i = 0; i < split.length; i++) { String cn = split[i].trim(); if (!cn.isEmpty() && !"class".equalsIgnoreCase(cn)) { try { classes.add(Class.forName(cn)); } catch (ClassNotFoundException e) { String msg = "Error registering class [" + cn + "] from [" + classNames + "]"; throw new RuntimeException(msg, e); } } } return classes; } private List getSearchJarsPackages(String searchPackages) { List hitList = new ArrayList(); if (searchPackages != null) { String[] entries = searchPackages.split("[ ,;]"); for (int i = 0; i < entries.length; i++) { hitList.add(entries[i].trim()); } } return hitList; } /** * Return the PersistBatch mode to use for 'batchOnCascade' taking into account if the database * platform supports getGeneratedKeys in batch mode. *

* Used to effectively turn off batchOnCascade for SQL Server - still allows explicit batch mode. *

*/ public PersistBatch appliedPersistBatchOnCascade() { return databasePlatform.isDisallowBatchOnCascade() ? PersistBatch.NONE : persistBatchOnCascade; } /** * Return the Jackson ObjectMapper. *

* Note that this is not strongly typed as Jackson ObjectMapper is an optional dependency. *

*/ public Object getObjectMapper() { return objectMapper; } /** * Set the Jackson ObjectMapper. *

* Note that this is not strongly typed as Jackson ObjectMapper is an optional dependency. *

*/ public void setObjectMapper(Object objectMapper) { this.objectMapper = objectMapper; } /** * Return true if eq("someProperty", null) should to generate "1=1" rather than "is null" sql expression. */ public boolean isExpressionEqualsWithNullAsNoop() { return expressionEqualsWithNullAsNoop; } /** * Set to true if you want eq("someProperty", null) to generate "1=1" rather than "is null" sql expression. *

* Setting this to true has the effect that eq(propertyName, value), ieq(propertyName, value) and * ne(propertyName, value) have no effect when the value is null. The expression factory adds a NoopExpression * which will add "1=1" into the SQL rather than "is null". *

*/ public void setExpressionEqualsWithNullAsNoop(boolean expressionEqualsWithNullAsNoop) { this.expressionEqualsWithNullAsNoop = expressionEqualsWithNullAsNoop; } /** * Return true if native ILIKE expression should be used if supported by the database platform (e.g. Postgres). */ public boolean isExpressionNativeIlike() { return expressionNativeIlike; } /** * Set to true to use native ILIKE expression if supported by the database platform (e.g. Postgres). */ public void setExpressionNativeIlike(boolean expressionNativeIlike) { this.expressionNativeIlike = expressionNativeIlike; } /** * Return true if L2 cache is disabled. */ public boolean isDisableL2Cache() { return disableL2Cache; } /** * Set to true to disable L2 caching. Typically useful in performance testing. */ public void setDisableL2Cache(boolean disableL2Cache) { this.disableL2Cache = disableL2Cache; } /** * Specify how UUID is stored. */ public enum DbUuid { /** * Store using native UUID in H2 and Postgres. */ AUTO, /** * Store using DB VARCHAR. */ VARCHAR, /** * Store using DB BINARY. */ BINARY } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy