org.jooq.meta.h2.H2Database Maven / Gradle / Ivy
/*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* https://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*
* Other licenses:
* -----------------------------------------------------------------------------
* Commercial licenses for this work are available. These replace the above
* Apache-2.0 license and offer limited warranties, support, maintenance, and
* commercial database integrations.
*
* For more information, please visit: https://www.jooq.org/legal/licensing
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*/
package org.jooq.meta.h2;
import static java.util.Arrays.asList;
import static org.jooq.Records.mapping;
import static org.jooq.SQLDialect.H2;
// ...
// ...
import static org.jooq.impl.DSL.arrayAgg;
import static org.jooq.impl.DSL.coalesce;
import static org.jooq.impl.DSL.concat;
import static org.jooq.impl.DSL.condition;
import static org.jooq.impl.DSL.falseCondition;
import static org.jooq.impl.DSL.field;
import static org.jooq.impl.DSL.inline;
import static org.jooq.impl.DSL.name;
import static org.jooq.impl.DSL.noCondition;
import static org.jooq.impl.DSL.not;
import static org.jooq.impl.DSL.nullif;
import static org.jooq.impl.DSL.nvl;
import static org.jooq.impl.DSL.one;
import static org.jooq.impl.DSL.row;
import static org.jooq.impl.DSL.select;
import static org.jooq.impl.DSL.upper;
import static org.jooq.impl.DSL.when;
import static org.jooq.impl.SQLDataType.BIGINT;
import static org.jooq.impl.SQLDataType.INTEGER;
import static org.jooq.impl.SQLDataType.NUMERIC;
import static org.jooq.impl.SQLDataType.VARCHAR;
import static org.jooq.meta.h2.information_schema.Tables.COLUMNS;
import static org.jooq.meta.h2.information_schema.Tables.CONSTRAINTS;
import static org.jooq.meta.h2.information_schema.Tables.CROSS_REFERENCES;
import static org.jooq.meta.h2.information_schema.Tables.DOMAINS;
import static org.jooq.meta.h2.information_schema.Tables.FUNCTION_ALIASES;
import static org.jooq.meta.h2.information_schema.Tables.INDEXES;
import static org.jooq.meta.h2.information_schema.Tables.SCHEMATA;
import static org.jooq.meta.h2.information_schema.Tables.SEQUENCES;
import static org.jooq.meta.h2.information_schema.Tables.TABLES;
import static org.jooq.meta.h2.information_schema.Tables.TYPE_INFO;
import static org.jooq.meta.h2.information_schema.Tables.VIEWS;
import static org.jooq.meta.hsqldb.information_schema.Tables.CHECK_CONSTRAINTS;
import static org.jooq.meta.hsqldb.information_schema.Tables.DOMAIN_CONSTRAINTS;
import static org.jooq.meta.hsqldb.information_schema.Tables.ELEMENT_TYPES;
import static org.jooq.meta.hsqldb.information_schema.Tables.KEY_COLUMN_USAGE;
import static org.jooq.meta.hsqldb.information_schema.Tables.REFERENTIAL_CONSTRAINTS;
import static org.jooq.meta.hsqldb.information_schema.Tables.ROUTINES;
import static org.jooq.meta.hsqldb.information_schema.Tables.TABLE_CONSTRAINTS;
import java.io.StringReader;
import java.math.BigDecimal;
import java.sql.SQLException;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.LinkedHashMap;
import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
import java.util.concurrent.atomic.AtomicBoolean;
import org.jooq.DSLContext;
import org.jooq.Field;
import org.jooq.Name;
import org.jooq.Record;
import org.jooq.Record12;
import org.jooq.Record14;
import org.jooq.Record4;
import org.jooq.Record5;
import org.jooq.Record6;
import org.jooq.Result;
import org.jooq.ResultQuery;
import org.jooq.SQLDialect;
import org.jooq.Select;
import org.jooq.SortOrder;
import org.jooq.Table;
import org.jooq.TableField;
// ...
// ...
import org.jooq.TableOptions.TableType;
import org.jooq.impl.DSL;
import org.jooq.meta.AbstractDatabase;
import org.jooq.meta.AbstractIndexDefinition;
import org.jooq.meta.ArrayDefinition;
import org.jooq.meta.CatalogDefinition;
import org.jooq.meta.ColumnDefinition;
import org.jooq.meta.DataTypeDefinition;
import org.jooq.meta.DefaultCheckConstraintDefinition;
import org.jooq.meta.DefaultDataTypeDefinition;
import org.jooq.meta.DefaultDomainDefinition;
import org.jooq.meta.DefaultEnumDefinition;
import org.jooq.meta.DefaultIndexColumnDefinition;
import org.jooq.meta.DefaultRelations;
import org.jooq.meta.DefaultSequenceDefinition;
import org.jooq.meta.DomainDefinition;
import org.jooq.meta.EnumDefinition;
import org.jooq.meta.IndexColumnDefinition;
import org.jooq.meta.IndexDefinition;
import org.jooq.meta.PackageDefinition;
import org.jooq.meta.ResultQueryDatabase;
import org.jooq.meta.RoutineDefinition;
import org.jooq.meta.SchemaDefinition;
import org.jooq.meta.SequenceDefinition;
import org.jooq.meta.TableDefinition;
// ...
import org.jooq.meta.UDTDefinition;
import org.jooq.meta.XMLSchemaCollectionDefinition;
import org.jooq.meta.hsqldb.information_schema.Tables;
import org.jooq.meta.hsqldb.information_schema.tables.CheckConstraints;
import org.jooq.meta.hsqldb.information_schema.tables.DomainConstraints;
import org.jooq.meta.hsqldb.information_schema.tables.Domains;
import org.jooq.meta.hsqldb.information_schema.tables.ElementTypes;
import org.jooq.meta.hsqldb.information_schema.tables.KeyColumnUsage;
import org.jooq.tools.StringUtils;
import org.jooq.tools.csv.CSVReader;
import org.jooq.util.h2.H2DataType;
import org.jetbrains.annotations.NotNull;
/**
* H2 implementation of {@link AbstractDatabase}
*
* @author Espen Stromsnes
*/
public class H2Database extends AbstractDatabase implements ResultQueryDatabase {
private static final long DEFAULT_SEQUENCE_CACHE = 32;
private static final long DEFAULT_SEQUENCE_MAXVALUE = Long.MAX_VALUE;
static final record ElementTypeLookupKey (String schema, String name, String identifier) {}
static final record ElementType (String dataType, Long length, Long precision, Long scale, String identifier, int dimension) {}
private Map ELEMENT_TYPE_LOOKUP;
ElementType elementTypeLookup(ElementTypeLookupKey key) {
if (ELEMENT_TYPE_LOOKUP == null) {
ElementTypes e = ELEMENT_TYPES;
ELEMENT_TYPE_LOOKUP = create().fetchMap(
select(
row(
e.OBJECT_SCHEMA,
e.OBJECT_NAME,
e.COLLECTION_TYPE_IDENTIFIER
).mapping(ElementTypeLookupKey::new),
row(
e.DATA_TYPE,
e.CHARACTER_MAXIMUM_LENGTH,
coalesce(e.DATETIME_PRECISION, e.NUMERIC_PRECISION),
e.NUMERIC_SCALE,
e.DTD_IDENTIFIER,
inline(1)
).mapping(ElementType::new)
)
.from(e)
.where(e.OBJECT_SCHEMA.in(getInputSchemata()))
);
AtomicBoolean repeat = new AtomicBoolean(true);
while (repeat.getAndSet(false)) {
ELEMENT_TYPE_LOOKUP.replaceAll((k, v) -> {
ElementType et = ELEMENT_TYPE_LOOKUP.get(new ElementTypeLookupKey(k.schema(), k.name(), v.identifier()));
if (et != null) {
repeat.set(true);
return new ElementType(et.dataType(), et.length(), et.precision(), et.scale(), et.identifier(), v.dimension() + 1); }
else
return v;
});
}
}
return ELEMENT_TYPE_LOOKUP.get(key);
}
@Override
protected DSLContext create0() {
return DSL.using(getConnection(), SQLDialect.H2);
}
@Override
protected boolean exists0(TableField, ?> field) {
return exists1(field, COLUMNS, COLUMNS.TABLE_SCHEMA, COLUMNS.TABLE_NAME, COLUMNS.COLUMN_NAME);
}
@Override
protected boolean exists0(Table> table) {
return exists1(table, TABLES, TABLES.TABLE_SCHEMA, TABLES.TABLE_NAME);
}
@Override
protected List getIndexes0() throws SQLException {
List result = new ArrayList<>();
// Same implementation as in HSQLDBDatabase and MySQLDatabase
Map> indexes = indexes()
.fetchGroups(
new Field[] {
INDEXES.TABLE_SCHEMA,
INDEXES.TABLE_NAME,
INDEXES.INDEX_NAME,
INDEXES.NON_UNIQUE
},
new Field[] {
INDEXES.COLUMN_NAME,
INDEXES.ORDINAL_POSITION
});
indexLoop:
for (Entry> entry : indexes.entrySet()) {
final Record index = entry.getKey();
final Result columns = entry.getValue();
final SchemaDefinition tableSchema = getSchema(index.get(INDEXES.TABLE_SCHEMA));
if (tableSchema == null)
continue indexLoop;
final String indexName = index.get(INDEXES.INDEX_NAME);
final String tableName = index.get(INDEXES.TABLE_NAME);
final TableDefinition table = getTable(tableSchema, tableName);
if (table == null)
continue indexLoop;
final boolean unique = !index.get(INDEXES.NON_UNIQUE, boolean.class);
// [#6310] [#6620] Function-based indexes are not yet supported
// [#16237] Alternatively, the column could be hidden or excluded
for (Record column : columns)
if (table.getColumn(column.get(INDEXES.COLUMN_NAME)) == null)
continue indexLoop;
result.add(new AbstractIndexDefinition(tableSchema, indexName, table, unique) {
List indexColumns = new ArrayList<>();
{
for (Record column : columns) {
indexColumns.add(new DefaultIndexColumnDefinition(
this,
table.getColumn(column.get(INDEXES.COLUMN_NAME)),
SortOrder.ASC,
column.get(INDEXES.ORDINAL_POSITION, int.class)
));
}
}
@Override
protected List getIndexColumns0() {
return indexColumns;
}
});
}
return result;
}
private ResultQuery> indexes() {
if (is2_0_202())
return indexes2_0();
else
return indexes1_4();
}
private ResultQuery> indexes2_0() {
return create()
.select(
INDEXES.TABLE_SCHEMA,
INDEXES.TABLE_NAME,
INDEXES.INDEX_NAME,
field(INDEXES.getQualifiedName().append("INDEX_TYPE_NAME")).ne(inline("UNIQUE INDEX")).as(INDEXES.NON_UNIQUE),
field("I.COLUMN_NAME", INDEXES.COLUMN_NAME).as(INDEXES.COLUMN_NAME),
field("I.ORDINAL_POSITION", INDEXES.ORDINAL_POSITION).as(INDEXES.ORDINAL_POSITION),
field("I.ORDERING_SPECIFICATION", INDEXES.ASC_OR_DESC).as(INDEXES.ASC_OR_DESC))
.from(INDEXES)
.join("INFORMATION_SCHEMA.INDEX_COLUMNS I")
.on("{0} = I.TABLE_SCHEMA", INDEXES.TABLE_SCHEMA)
.and("{0} = I.TABLE_NAME", INDEXES.TABLE_NAME)
.and("{0} = I.INDEX_NAME", INDEXES.INDEX_NAME)
.where(INDEXES.TABLE_SCHEMA.in(getInputSchemata()))
.and(getIncludeSystemIndexes()
? noCondition()
: not(condition(INDEXES.IS_GENERATED)))
.orderBy(
INDEXES.TABLE_SCHEMA,
INDEXES.TABLE_NAME,
INDEXES.INDEX_NAME,
field("I.ORDINAL_POSITION"));
}
private ResultQuery> indexes1_4() {
return create()
.select(
INDEXES.TABLE_SCHEMA,
INDEXES.TABLE_NAME,
INDEXES.INDEX_NAME,
INDEXES.NON_UNIQUE,
INDEXES.COLUMN_NAME,
INDEXES.ORDINAL_POSITION,
INDEXES.ASC_OR_DESC)
.from(INDEXES)
.where(INDEXES.TABLE_SCHEMA.in(getInputSchemata()))
.and(getIncludeSystemIndexes()
? noCondition()
: not(condition(INDEXES.IS_GENERATED)))
.orderBy(
INDEXES.TABLE_SCHEMA,
INDEXES.TABLE_NAME,
INDEXES.INDEX_NAME,
INDEXES.ORDINAL_POSITION);
}
@Override
protected void loadPrimaryKeys(DefaultRelations relations) throws SQLException {
for (Record record : primaryKeys(getInputSchemata())) {
SchemaDefinition schema = getSchema(record.get(CONSTRAINTS.TABLE_SCHEMA));
if (schema != null) {
String tableName = record.get(CONSTRAINTS.TABLE_NAME);
String primaryKey = record.get(CONSTRAINTS.CONSTRAINT_NAME);
String columnName = record.get(INDEXES.COLUMN_NAME);
TableDefinition table = getTable(schema, tableName);
if (table != null)
relations.addPrimaryKey(primaryKey, table, table.getColumn(columnName));
}
}
}
@Override
protected void loadUniqueKeys(DefaultRelations relations) throws SQLException {
for (Record record : uniqueKeys(getInputSchemata())) {
SchemaDefinition schema = getSchema(record.get(CONSTRAINTS.TABLE_SCHEMA));
if (schema != null) {
String tableName = record.get(CONSTRAINTS.TABLE_NAME);
String primaryKey = record.get(CONSTRAINTS.CONSTRAINT_NAME);
String columnName = record.get(INDEXES.COLUMN_NAME);
TableDefinition table = getTable(schema, tableName);
if (table != null)
relations.addUniqueKey(primaryKey, table, table.getColumn(columnName));
}
}
}
@Override
public ResultQuery> primaryKeys(List schemas) {
if (is2_0_202())
return keys2_0(schemas, Arrays.>asList(inline("PRIMARY KEY"), inline("PRIMARY_KEY")));
// Workaround for https://github.com/h2database/h2database/issues/1000
else
return keys1_4(schemas, Arrays.>asList(inline("PRIMARY KEY"), inline("PRIMARY_KEY")));
}
@Override
public ResultQuery> uniqueKeys(List schemas) {
if (is2_0_202())
return keys2_0(schemas, Arrays.>asList(inline("UNIQUE")));
else
return keys1_4(schemas, Arrays.>asList(inline("UNIQUE")));
}
private ResultQuery> keys2_0(List schemas, List> constraintTypes) {
return create()
.select(
KEY_COLUMN_USAGE.TABLE_CATALOG,
KEY_COLUMN_USAGE.TABLE_SCHEMA,
KEY_COLUMN_USAGE.TABLE_NAME,
KEY_COLUMN_USAGE.CONSTRAINT_NAME,
KEY_COLUMN_USAGE.COLUMN_NAME,
KEY_COLUMN_USAGE.ORDINAL_POSITION.coerce(INTEGER))
.from(KEY_COLUMN_USAGE)
.where(KEY_COLUMN_USAGE.tableConstraints().CONSTRAINT_TYPE.in(constraintTypes))
.and(KEY_COLUMN_USAGE.tableConstraints().TABLE_SCHEMA.in(schemas))
.orderBy(
KEY_COLUMN_USAGE.TABLE_SCHEMA.asc(),
KEY_COLUMN_USAGE.TABLE_NAME.asc(),
KEY_COLUMN_USAGE.CONSTRAINT_NAME.asc(),
KEY_COLUMN_USAGE.ORDINAL_POSITION.asc());
}
private ResultQuery> keys1_4(List schemas, List> constraintTypes) {
return create().select(
CONSTRAINTS.TABLE_CATALOG,
CONSTRAINTS.TABLE_SCHEMA,
CONSTRAINTS.TABLE_NAME,
CONSTRAINTS.CONSTRAINT_NAME,
INDEXES.COLUMN_NAME,
INDEXES.ORDINAL_POSITION.coerce(INTEGER))
.from(CONSTRAINTS)
.join(INDEXES)
.on(CONSTRAINTS.TABLE_SCHEMA.eq(INDEXES.TABLE_SCHEMA))
.and(CONSTRAINTS.TABLE_NAME.eq(INDEXES.TABLE_NAME))
.and(CONSTRAINTS.UNIQUE_INDEX_NAME.eq(INDEXES.INDEX_NAME))
.where(CONSTRAINTS.TABLE_SCHEMA.in(schemas))
.and(CONSTRAINTS.CONSTRAINT_TYPE.in(constraintTypes))
.orderBy(
CONSTRAINTS.TABLE_SCHEMA,
CONSTRAINTS.CONSTRAINT_NAME,
INDEXES.ORDINAL_POSITION);
}
@Override
protected void loadForeignKeys(DefaultRelations relations) throws SQLException {
if (is2_0_202())
loadForeignKeys2_0(relations);
else
loadForeignKeys1_4(relations);
}
private void loadForeignKeys2_0(DefaultRelations relations) {
KeyColumnUsage fkKcu = KEY_COLUMN_USAGE.as("fk_kcu");
KeyColumnUsage pkKcu = KEY_COLUMN_USAGE.as("pk_kcu");
for (Record record : create()
.select(
REFERENTIAL_CONSTRAINTS.UNIQUE_CONSTRAINT_NAME,
REFERENTIAL_CONSTRAINTS.UNIQUE_CONSTRAINT_SCHEMA,
TABLE_CONSTRAINTS.TABLE_NAME,
fkKcu.CONSTRAINT_NAME,
fkKcu.TABLE_SCHEMA,
fkKcu.TABLE_NAME,
fkKcu.COLUMN_NAME,
pkKcu.COLUMN_NAME
)
.from(REFERENTIAL_CONSTRAINTS)
.join(fkKcu)
.on(fkKcu.CONSTRAINT_SCHEMA.equal(REFERENTIAL_CONSTRAINTS.CONSTRAINT_SCHEMA))
.and(fkKcu.CONSTRAINT_NAME.equal(REFERENTIAL_CONSTRAINTS.CONSTRAINT_NAME))
.join(TABLE_CONSTRAINTS)
.on(TABLE_CONSTRAINTS.CONSTRAINT_SCHEMA.eq(REFERENTIAL_CONSTRAINTS.UNIQUE_CONSTRAINT_SCHEMA))
.and(TABLE_CONSTRAINTS.CONSTRAINT_NAME.eq(REFERENTIAL_CONSTRAINTS.UNIQUE_CONSTRAINT_NAME))
.join(pkKcu)
.on(pkKcu.CONSTRAINT_SCHEMA.eq(TABLE_CONSTRAINTS.CONSTRAINT_SCHEMA))
.and(pkKcu.CONSTRAINT_NAME.eq(TABLE_CONSTRAINTS.CONSTRAINT_NAME))
.and(pkKcu.ORDINAL_POSITION.eq(fkKcu.POSITION_IN_UNIQUE_CONSTRAINT))
.where(fkKcu.TABLE_SCHEMA.in(getInputSchemata()))
.orderBy(
fkKcu.TABLE_SCHEMA.asc(),
fkKcu.TABLE_NAME.asc(),
fkKcu.CONSTRAINT_NAME.asc(),
fkKcu.ORDINAL_POSITION.asc())
) {
SchemaDefinition foreignKeySchema = getSchema(record.get(fkKcu.TABLE_SCHEMA));
SchemaDefinition uniqueKeySchema = getSchema(record.get(REFERENTIAL_CONSTRAINTS.UNIQUE_CONSTRAINT_SCHEMA));
String foreignKey = record.get(fkKcu.CONSTRAINT_NAME);
String foreignKeyTableName = record.get(fkKcu.TABLE_NAME);
String foreignKeyColumn = record.get(fkKcu.COLUMN_NAME);
String uniqueKey = record.get(REFERENTIAL_CONSTRAINTS.UNIQUE_CONSTRAINT_NAME);
String uniqueKeyTableName = record.get(TABLE_CONSTRAINTS.TABLE_NAME);
String uniqueKeyColumn = record.get(pkKcu.COLUMN_NAME);
TableDefinition foreignKeyTable = getTable(foreignKeySchema, foreignKeyTableName);
TableDefinition uniqueKeyTable = getTable(uniqueKeySchema, uniqueKeyTableName);
if (foreignKeyTable != null && uniqueKeyTable != null)
relations.addForeignKey(
foreignKey,
foreignKeyTable,
foreignKeyTable.getColumn(foreignKeyColumn),
uniqueKey,
uniqueKeyTable,
uniqueKeyTable.getColumn(uniqueKeyColumn),
true
);
}
}
private void loadForeignKeys1_4(DefaultRelations relations) {
for (Record record : create().select(
CROSS_REFERENCES.FK_NAME,
CROSS_REFERENCES.FKTABLE_NAME,
CROSS_REFERENCES.FKTABLE_SCHEMA,
CROSS_REFERENCES.FKCOLUMN_NAME,
CROSS_REFERENCES.PKCOLUMN_NAME,
CROSS_REFERENCES.referencedConstraint().CONSTRAINT_NAME,
CROSS_REFERENCES.referencedConstraint().TABLE_NAME,
CROSS_REFERENCES.referencedConstraint().CONSTRAINT_SCHEMA)
.from(CROSS_REFERENCES)
.where(CROSS_REFERENCES.FKTABLE_SCHEMA.in(getInputSchemata()))
// Workaround for https://github.com/h2database/h2database/issues/1000
.and(CROSS_REFERENCES.referencedConstraint().CONSTRAINT_TYPE.in("PRIMARY KEY", "PRIMARY_KEY", "UNIQUE"))
.orderBy(
CROSS_REFERENCES.FKTABLE_SCHEMA.asc(),
CROSS_REFERENCES.FK_NAME.asc(),
CROSS_REFERENCES.ORDINAL_POSITION.asc())
.fetch()) {
SchemaDefinition foreignKeySchema = getSchema(record.get(CROSS_REFERENCES.FKTABLE_SCHEMA));
SchemaDefinition uniqueKeySchema = getSchema(record.get(CROSS_REFERENCES.referencedConstraint().CONSTRAINT_SCHEMA));
if (foreignKeySchema != null && uniqueKeySchema != null) {
String foreignKey = record.get(CROSS_REFERENCES.FK_NAME);
String foreignKeyTableName = record.get(CROSS_REFERENCES.FKTABLE_NAME);
String foreignKeyColumn = record.get(CROSS_REFERENCES.FKCOLUMN_NAME);
String uniqueKey = record.get(CROSS_REFERENCES.referencedConstraint().CONSTRAINT_NAME);
String uniqueKeyTableName = record.get(CROSS_REFERENCES.referencedConstraint().TABLE_NAME);
String uniqueKeyColumn = record.get(CROSS_REFERENCES.PKCOLUMN_NAME);
TableDefinition foreignKeyTable = getTable(foreignKeySchema, foreignKeyTableName);
TableDefinition uniqueKeyTable = getTable(uniqueKeySchema, uniqueKeyTableName);
if (foreignKeyTable != null && uniqueKeyTable != null)
relations.addForeignKey(
foreignKey,
foreignKeyTable,
foreignKeyTable.getColumn(foreignKeyColumn),
uniqueKey,
uniqueKeyTable,
uniqueKeyTable.getColumn(uniqueKeyColumn),
true
);
}
}
}
@Override
protected void loadCheckConstraints(DefaultRelations relations) throws SQLException {
if (is2_0_202())
loadCheckConstraints2_0(relations);
else
loadCheckConstraints1_4(relations);
}
private void loadCheckConstraints2_0(DefaultRelations relations) {
CheckConstraints cc = CHECK_CONSTRAINTS.as("cc");
for (Record record : create()
.select(
cc.tableConstraints().TABLE_SCHEMA,
cc.tableConstraints().TABLE_NAME,
cc.CONSTRAINT_NAME,
cc.CHECK_CLAUSE
)
.from(cc)
.where(cc.tableConstraints().TABLE_SCHEMA.in(getInputSchemata()))
) {
SchemaDefinition schema = getSchema(record.get(cc.tableConstraints().TABLE_SCHEMA));
TableDefinition table = getTable(schema, record.get(cc.tableConstraints().TABLE_NAME));
if (table != null) {
relations.addCheckConstraint(table, new DefaultCheckConstraintDefinition(
schema,
table,
record.get(cc.CONSTRAINT_NAME),
record.get(cc.CHECK_CLAUSE)
));
}
}
}
private void loadCheckConstraints1_4(DefaultRelations relations) {
// TODO [https://github.com/h2database/h2database/issues/2286]
// Starting from H2 1.4.201, we should no longer produce the below UNION
Select> inlineChecks = is2_0_202()
? select(inline(""), inline(""), inline(""), inline("")).where(falseCondition())
: select(
COLUMNS.TABLE_SCHEMA,
COLUMNS.TABLE_NAME,
COLUMNS.CHECK_CONSTRAINT,
COLUMNS.CHECK_CONSTRAINT
)
.from(COLUMNS)
.where(nvl(COLUMNS.CHECK_CONSTRAINT, inline("")).ne(inline("")))
.and(COLUMNS.TABLE_SCHEMA.in(getInputSchemata()));
for (Record record : create()
.select(
CONSTRAINTS.TABLE_SCHEMA,
CONSTRAINTS.TABLE_NAME,
CONSTRAINTS.CONSTRAINT_NAME,
CONSTRAINTS.CHECK_EXPRESSION
)
.from(CONSTRAINTS)
.where(CONSTRAINTS.CONSTRAINT_TYPE.eq("CHECK"))
.and(CONSTRAINTS.TABLE_SCHEMA.in(getInputSchemata()))
.union(inlineChecks)) {
SchemaDefinition schema = getSchema(record.get(CONSTRAINTS.TABLE_SCHEMA));
if (schema != null) {
TableDefinition table = getTable(schema, record.get(CONSTRAINTS.TABLE_NAME));
if (table != null) {
relations.addCheckConstraint(table, new DefaultCheckConstraintDefinition(
schema,
table,
record.get(CONSTRAINTS.CONSTRAINT_NAME),
record.get(CONSTRAINTS.CHECK_EXPRESSION)
));
}
}
}
}
@Override
protected List getCatalogs0() throws SQLException {
List result = new ArrayList<>();
result.add(new CatalogDefinition(this, "", ""));
return result;
}
@Override
protected List getSchemata0() throws SQLException {
return
create().select(SCHEMATA.SCHEMA_NAME, SCHEMATA.REMARKS)
.from(SCHEMATA)
.fetch(mapping((s, r) -> new SchemaDefinition(this, s, r)));
}
@Override
public ResultQuery> sources(List schemas) {
return create()
.select(
VIEWS.TABLE_CATALOG,
VIEWS.TABLE_SCHEMA,
VIEWS.TABLE_NAME,
concat(inline("create view \""), VIEWS.TABLE_NAME, inline("\" as "), VIEWS.VIEW_DEFINITION).as(VIEWS.VIEW_DEFINITION))
.from(VIEWS)
.where(VIEWS.TABLE_SCHEMA.in(schemas))
.orderBy(
VIEWS.TABLE_SCHEMA,
VIEWS.TABLE_NAME)
;
}
@Override
public ResultQuery> comments(List schemas) {
Table> c =
select(
TABLES.TABLE_CATALOG,
TABLES.TABLE_SCHEMA,
TABLES.TABLE_NAME,
inline(null, VARCHAR).as(COLUMNS.COLUMN_NAME),
TABLES.REMARKS)
.from(TABLES)
.where(TABLES.REMARKS.isNotNull())
.unionAll(
select(
COLUMNS.TABLE_CATALOG,
COLUMNS.TABLE_SCHEMA,
COLUMNS.TABLE_NAME,
COLUMNS.COLUMN_NAME,
COLUMNS.REMARKS)
.from(COLUMNS)
.where(COLUMNS.REMARKS.isNotNull()))
.asTable("c");
return create()
.select(
c.field(TABLES.TABLE_CATALOG),
c.field(TABLES.TABLE_SCHEMA),
c.field(TABLES.TABLE_NAME),
c.field(COLUMNS.COLUMN_NAME),
c.field(TABLES.REMARKS))
.from(c)
.where(c.field(TABLES.TABLE_SCHEMA).in(schemas))
.orderBy(1, 2, 3, 4);
}
@Override
public ResultQuery> sequences(List schemas) {
Field minValue = is2_0_202()
? field(SEQUENCES.getQualifiedName().append("MINIMUM_VALUE"), SEQUENCES.MIN_VALUE.getDataType())
: SEQUENCES.MIN_VALUE;
Field maxValue = is2_0_202()
? field(SEQUENCES.getQualifiedName().append("MAXIMUM_VALUE"), SEQUENCES.MAX_VALUE.getDataType())
: SEQUENCES.MAX_VALUE;
Field isCycle = is2_0_202()
? field(SEQUENCES.getQualifiedName().append("CYCLE_OPTION")).eq(inline("YES"))
: SEQUENCES.IS_CYCLE;
return create()
.select(
inline(null, VARCHAR).as("catalog"),
SEQUENCES.SEQUENCE_SCHEMA,
SEQUENCES.SEQUENCE_NAME,
inline("BIGINT").as("type_name"),
inline(null, INTEGER).as("precision"),
inline(null, INTEGER).as("scale"),
inline(null, BIGINT).as("start_value"),
nullif(SEQUENCES.INCREMENT, inline(1L)).as(SEQUENCES.INCREMENT),
nullif(minValue, inline(1L)).coerce(NUMERIC).as(SEQUENCES.MIN_VALUE),
nullif(maxValue, inline(DEFAULT_SEQUENCE_MAXVALUE)).coerce(NUMERIC).as(SEQUENCES.MAX_VALUE),
isCycle.as(SEQUENCES.IS_CYCLE),
nullif(SEQUENCES.CACHE, inline(DEFAULT_SEQUENCE_CACHE)).as(SEQUENCES.CACHE)
)
.from(SEQUENCES)
.where(SEQUENCES.SEQUENCE_SCHEMA.in(schemas))
.and(!getIncludeSystemSequences() ? upper(SEQUENCES.SEQUENCE_NAME).notLike(inline("SYSTEM!_SEQUENCE!_%"), '!') : noCondition())
.orderBy(
SEQUENCES.SEQUENCE_SCHEMA,
SEQUENCES.SEQUENCE_NAME);
}
@Override
protected List getSequences0() throws SQLException {
List result = new ArrayList<>();
for (Record record : sequences(getInputSchemata())) {
SchemaDefinition schema = getSchema(record.get(SEQUENCES.SEQUENCE_SCHEMA));
if (schema != null) {
String name = record.get(SEQUENCES.SEQUENCE_NAME);
DefaultDataTypeDefinition type = new DefaultDataTypeDefinition(
this,
schema,
H2DataType.BIGINT.getTypeName()
);
result.add(new DefaultSequenceDefinition(
schema,
name,
type,
null,
null, // H2 doesn't support Postgres-style START WITH
record.get(SEQUENCES.INCREMENT),
record.get(SEQUENCES.MIN_VALUE),
record.get(SEQUENCES.MAX_VALUE),
record.get(SEQUENCES.IS_CYCLE),
record.get(SEQUENCES.CACHE)));
}
}
return result;
}
@Override
public ResultQuery> enums(List schemas) {
return null;
}
static record TableRecord(String schema, String table, TableType type, String comment) {}
@Override
protected List getTables0() throws SQLException {
List result = new ArrayList<>();
for (TableRecord r : create().select(
TABLES.TABLE_SCHEMA,
TABLES.TABLE_NAME,
when(TABLES.TABLE_TYPE.eq(inline("VIEW")), inline(TableType.VIEW.name()))
.when(TABLES.STORAGE_TYPE.like(inline("%TEMPORARY%")), inline(TableType.TEMPORARY.name()))
.else_(inline(TableType.TABLE.name())).convertFrom(TableType::valueOf).as("table_type"),
TABLES.REMARKS)
.from(TABLES)
.where(TABLES.TABLE_SCHEMA.in(getInputSchemata()))
.orderBy(
TABLES.TABLE_SCHEMA,
TABLES.TABLE_NAME)
.fetch(mapping(TableRecord::new))) {
SchemaDefinition schema = getSchema(r.schema);
if (schema != null)
result.add(new H2TableDefinition(schema, r.table, r.comment, r.type, null));
}
return result;
}
@Override
protected List getRoutines0() throws SQLException {
if (is2_0_202())
return getRoutines2_0();
else
return getRoutines1_4();
}
private List getRoutines2_0() {
List result = new ArrayList<>();
for (Record record : create()
.select(
ROUTINES.ROUTINE_SCHEMA,
ROUTINES.ROUTINE_NAME,
ROUTINES.SPECIFIC_NAME,
coalesce(
concat(ELEMENT_TYPES.DATA_TYPE, inline(" ARRAY")),
ROUTINES.DATA_TYPE,
when(ROUTINES.ROUTINE_TYPE.eq(inline("FUNCTION")), inline("RESULT_SET"))
).as(ROUTINES.DATA_TYPE),
ROUTINES.NUMERIC_PRECISION,
ROUTINES.NUMERIC_SCALE)
.from(ROUTINES)
.leftOuterJoin(ELEMENT_TYPES)
.on(ROUTINES.ROUTINE_SCHEMA.equal(ELEMENT_TYPES.OBJECT_SCHEMA))
.and(ROUTINES.ROUTINE_NAME.equal(ELEMENT_TYPES.OBJECT_NAME))
.and(ROUTINES.DTD_IDENTIFIER.equal(ELEMENT_TYPES.COLLECTION_TYPE_IDENTIFIER))
.where(ROUTINES.ROUTINE_SCHEMA.in(getInputSchemata()))
.and(tableValuedFunctions()
? ROUTINES.DATA_TYPE.isNull().or(ROUTINES.DATA_TYPE.notLike(inline("ROW(%")))
: noCondition())
.orderBy(
ROUTINES.ROUTINE_SCHEMA,
ROUTINES.ROUTINE_NAME)
) {
result.add(new H2RoutineDefinition(
getSchema(record.get(ROUTINES.ROUTINE_SCHEMA)),
record.get(ROUTINES.ROUTINE_NAME),
record.get(ROUTINES.SPECIFIC_NAME),
record.get(ROUTINES.DATA_TYPE),
record.get(ROUTINES.NUMERIC_PRECISION),
record.get(ROUTINES.NUMERIC_SCALE),
null,
null
));
}
return result;
}
private List getRoutines1_4() {
List result = new ArrayList<>();
Field overloaded = field(select(DSL.exists(
select(one())
.from(FUNCTION_ALIASES.as("a"))
.where(field(name("a", FUNCTION_ALIASES.ALIAS_SCHEMA.getName())).eq(FUNCTION_ALIASES.ALIAS_SCHEMA))
.and(field(name("a", FUNCTION_ALIASES.ALIAS_NAME.getName())).eq(FUNCTION_ALIASES.ALIAS_NAME))
.and(field(name("a", FUNCTION_ALIASES.COLUMN_COUNT.getName())).ne(FUNCTION_ALIASES.COLUMN_COUNT))
))).as("overloaded");
for (Record record : create()
.select(
FUNCTION_ALIASES.ALIAS_SCHEMA,
FUNCTION_ALIASES.ALIAS_NAME,
FUNCTION_ALIASES.REMARKS,
FUNCTION_ALIASES.DATA_TYPE,
FUNCTION_ALIASES.RETURNS_RESULT,
FUNCTION_ALIASES.COLUMN_COUNT,
overloaded,
TYPE_INFO.TYPE_NAME,
TYPE_INFO.PRECISION,
TYPE_INFO.MAXIMUM_SCALE)
.from(FUNCTION_ALIASES)
.leftOuterJoin(TYPE_INFO)
.on(FUNCTION_ALIASES.DATA_TYPE.equal(TYPE_INFO.DATA_TYPE))
// [#1256] TODO implement this correctly. Not sure if POS = 0
// is the right predicate to rule out duplicate entries in TYPE_INFO
.and(TYPE_INFO.POS.equal(0))
.where(FUNCTION_ALIASES.ALIAS_SCHEMA.in(getInputSchemata()))
.and(FUNCTION_ALIASES.RETURNS_RESULT.in((short) 1, (short) 2))
.orderBy(FUNCTION_ALIASES.ALIAS_NAME).fetch()) {
SchemaDefinition schema = getSchema(record.get(FUNCTION_ALIASES.ALIAS_SCHEMA));
if (schema != null) {
String name = record.get(FUNCTION_ALIASES.ALIAS_NAME);
String comment = record.get(FUNCTION_ALIASES.REMARKS);
String typeName = record.get(TYPE_INFO.TYPE_NAME);
Integer precision = record.get(TYPE_INFO.PRECISION);
Short scale = record.get(TYPE_INFO.MAXIMUM_SCALE);
String overload = record.get(overloaded)
? record.get(FUNCTION_ALIASES.COLUMN_COUNT, String.class)
: null;
result.add(new H2RoutineDefinition(schema, name, null, typeName, precision, scale, comment, overload));
}
}
return result;
}
@Override
protected List getPackages0() throws SQLException {
List result = new ArrayList<>();
return result;
}
@Override
protected List getEnums0() throws SQLException {
List result = new ArrayList<>();
if (!is1_4_197() && !is2_0_202())
return result;
if (is2_0_202()) {
getInlineEnums2_0(result);
getDomainEnums2_0(result);
}
else {
getInlineEnums1_4(result);
getDomainEnums1_4(result);
}
return result;
}
private void getInlineEnums2_0(List result) {
// TODO: Re-generate the H2 schema and use generated code to run this query
create().resultQuery("""
select c.TABLE_SCHEMA , c.TABLE_NAME , c.COLUMN_NAME , array_agg(ev.value_name order by ev.VALUE_ORDINAL)
from INFORMATION_SCHEMA.ENUM_VALUES ev
join INFORMATION_SCHEMA.COLUMNS c
on ev.OBJECT_SCHEMA = c.TABLE_SCHEMA
and ev.OBJECT_NAME = c.TABLE_NAME
and ev.ENUM_IDENTIFIER = c.DTD_IDENTIFIER
where ev.OBJECT_TYPE = 'TABLE'
and c.DOMAIN_NAME IS NULL
group by c.TABLE_SCHEMA , c.TABLE_NAME , c.COLUMN_NAME
order by 1, 2, 3
""")
.coerce(COLUMNS.TABLE_SCHEMA, COLUMNS.TABLE_NAME, COLUMNS.COLUMN_NAME, arrayAgg(COLUMNS.COLUMN_NAME))
.forEach(r -> {
SchemaDefinition schema = getSchema(r.value1());
if (schema == null)
return;
// [#1237] Don't generate enum classes for columns in MySQL tables
// that are excluded from code generation
TableDefinition tableDefinition = getTable(schema, r.value2());
if (tableDefinition != null) {
ColumnDefinition columnDefinition = tableDefinition.getColumn(r.value3());
if (columnDefinition != null) {
// [#1137] Avoid generating enum classes for enum types that
// are explicitly forced to another type
if (getConfiguredForcedType(columnDefinition, columnDefinition.getType()) == null) {
String name = r.value2() + "_" + r.value3();
DefaultEnumDefinition definition = new DefaultEnumDefinition(schema, name, "", true);
for (String string : r.value4())
definition.addLiteral(string);
result.add(definition);
}
}
}
});
}
static record EnumRecord (String schema, String table, String column, String type) {}
private void getInlineEnums1_4(List result) {
enumLoop:
for (EnumRecord r : create()
.select(
COLUMNS.TABLE_SCHEMA,
COLUMNS.TABLE_NAME,
COLUMNS.COLUMN_NAME,
COLUMNS.COLUMN_TYPE)
.from(COLUMNS)
.where(
COLUMNS.COLUMN_TYPE.like("ENUM(%)%").and(
COLUMNS.TABLE_SCHEMA.in(getInputSchemata())))
.orderBy(
COLUMNS.TABLE_SCHEMA.asc(),
COLUMNS.TABLE_NAME.asc(),
COLUMNS.COLUMN_NAME.asc())
.fetch(mapping(EnumRecord::new))) {
SchemaDefinition schema = getSchema(r.schema);
if (schema == null)
continue enumLoop;
// [#1237] Don't generate enum classes for columns in MySQL tables
// that are excluded from code generation
TableDefinition tableDefinition = getTable(schema, r.table);
if (tableDefinition != null) {
ColumnDefinition columnDefinition = tableDefinition.getColumn(r.column);
if (columnDefinition != null) {
// [#1137] Avoid generating enum classes for enum types that
// are explicitly forced to another type
if (getConfiguredForcedType(columnDefinition, columnDefinition.getType()) == null) {
String name = r.table + "_" + r.column;
DefaultEnumDefinition definition = new DefaultEnumDefinition(schema, name, "", true);
CSVReader reader = new CSVReader(
new StringReader(r.type.replaceAll("(^enum\\()|(\\)[^)]*$)", ""))
,',' // Separator
,'\'' // Quote character
,true // Strict quotes
);
for (String string : reader.next())
definition.addLiteral(string);
result.add(definition);
}
}
}
}
}
private void getDomainEnums2_0(List result) {
// TODO: Re-generate the H2 schema and use generated code to run this query
create().resultQuery("""
select object_schema, object_name, array_agg(ev.value_name order by ev.VALUE_ORDINAL)
from INFORMATION_SCHEMA.ENUM_VALUES ev
where ev.OBJECT_TYPE = 'DOMAIN'
group by object_schema, object_name
order by 1, 2
""")
.coerce(COLUMNS.TABLE_SCHEMA, COLUMNS.TABLE_NAME, arrayAgg(COLUMNS.COLUMN_NAME))
.forEach(r -> {
SchemaDefinition schema = getSchema(r.value1());
if (schema == null)
return;
DefaultEnumDefinition definition = new DefaultEnumDefinition(schema, r.value2(), "");
for (String string : r.value3())
definition.addLiteral(string);
result.add(definition);
});
}
static record DomainRecord (String schema, String name, String sql) {}
private void getDomainEnums1_4(List result) {
enumLoop:
for (DomainRecord r : create()
.select(
DOMAINS.DOMAIN_SCHEMA,
DOMAINS.DOMAIN_NAME,
DOMAINS.SQL)
.from(DOMAINS)
.where(DOMAINS.TYPE_NAME.eq(inline("ENUM")))
.and(DOMAINS.DOMAIN_SCHEMA.in(getInputSchemata()))
.orderBy(
DOMAINS.DOMAIN_SCHEMA,
DOMAINS.DOMAIN_NAME)
.fetch(mapping(DomainRecord::new))) {
SchemaDefinition schema = getSchema(r.schema);
if (schema == null)
continue enumLoop;
DefaultEnumDefinition definition = new DefaultEnumDefinition(schema, r.name, "");
CSVReader reader = new CSVReader(
new StringReader(r.sql.replaceAll("(?i:(^.*as enum\\()|(\\)[^)]*$))", ""))
,',' // Separator
,'\'' // Quote character
,true // Strict quotes
);
for (String string : reader.next())
definition.addLiteral(string);
result.add(definition);
}
}
@Override
protected List getDomains0() throws SQLException {
if (is2_0_202())
return getDomains2_0();
else
return getDomains1_4();
}
private List getDomains2_0() {
Map result = new LinkedHashMap<>();
Domains d = Tables.DOMAINS.as("d");
DomainConstraints dc = DOMAIN_CONSTRAINTS.as("dc");
for (Record record : create()
.select(
d.DOMAIN_SCHEMA,
d.DOMAIN_NAME,
d.DATA_TYPE,
d.CHARACTER_MAXIMUM_LENGTH,
coalesce(d.NUMERIC_PRECISION, d.DATETIME_PRECISION).as(d.NUMERIC_PRECISION),
d.NUMERIC_SCALE,
d.DOMAIN_DEFAULT,
dc.checkConstraints().CHECK_CLAUSE)
.from(d)
.leftJoin(dc)
.on(d.DOMAIN_CATALOG.eq(dc.DOMAIN_CATALOG))
.and(d.DOMAIN_SCHEMA.eq(dc.DOMAIN_SCHEMA))
.and(d.DOMAIN_NAME.eq(dc.DOMAIN_NAME))
.where(d.DOMAIN_SCHEMA.in(getInputSchemata()))
.and(d.DATA_TYPE.ne(inline("ENUM")))
.orderBy(
d.DOMAIN_SCHEMA,
d.DOMAIN_NAME,
dc.checkConstraints().CONSTRAINT_NAME
)
) {
String schemaName = record.get(d.DOMAIN_SCHEMA);
String domainName = record.get(d.DOMAIN_NAME);
String check = record.get(dc.checkConstraints().CHECK_CLAUSE);
DefaultDomainDefinition domain = result.computeIfAbsent(name(schemaName, domainName), k -> {
SchemaDefinition schema = getSchema(schemaName);
DataTypeDefinition baseType = new DefaultDataTypeDefinition(
this,
schema,
record.get(d.DATA_TYPE),
record.get(d.CHARACTER_MAXIMUM_LENGTH),
record.get(d.NUMERIC_PRECISION),
record.get(d.NUMERIC_SCALE),
true,
record.get(d.DOMAIN_DEFAULT)
);
return new DefaultDomainDefinition(
schema,
domainName,
baseType
);
});
if (!StringUtils.isBlank(check))
domain.addCheckClause(check);
}
return new ArrayList<>(result.values());
}
private List getDomains1_4() {
List result = new ArrayList<>();
for (Record record : create()
.select(
DOMAINS.DOMAIN_SCHEMA,
DOMAINS.DOMAIN_NAME,
DOMAINS.TYPE_NAME,
DOMAINS.PRECISION,
DOMAINS.SCALE,
DOMAINS.IS_NULLABLE,
DOMAINS.COLUMN_DEFAULT,
DOMAINS.CHECK_CONSTRAINT)
.from(DOMAINS)
.where(DOMAINS.DOMAIN_SCHEMA.in(getInputSchemata()))
// [#7917] Starting from 1.4.198, ENUM types are stored as domains
.and(DOMAINS.TYPE_NAME.ne(inline("ENUM")))
.orderBy(DOMAINS.DOMAIN_SCHEMA, DOMAINS.DOMAIN_NAME)) {
// [#7644] [#11721] H2 puts DATETIME_PRECISION in NUMERIC_SCALE column
boolean isTimestamp = record.get(DOMAINS.TYPE_NAME).trim().toLowerCase().startsWith("timestamp");
SchemaDefinition schema = getSchema(record.get(DOMAINS.DOMAIN_SCHEMA));
DataTypeDefinition baseType = new DefaultDataTypeDefinition(
this,
schema,
record.get(DOMAINS.TYPE_NAME),
record.get(DOMAINS.PRECISION),
isTimestamp
? record.get(DOMAINS.SCALE)
: record.get(DOMAINS.PRECISION),
isTimestamp
? 0
: record.get(DOMAINS.SCALE),
!record.get(DOMAINS.IS_NULLABLE, boolean.class),
record.get(DOMAINS.COLUMN_DEFAULT)
);
DefaultDomainDefinition domain = new DefaultDomainDefinition(
schema,
record.get(DOMAINS.DOMAIN_NAME),
baseType
);
if (!StringUtils.isBlank(record.get(DOMAINS.CHECK_CONSTRAINT)))
domain.addCheckClause(record.get(DOMAINS.CHECK_CONSTRAINT));
result.add(domain);
}
return result;
}
@Override
protected List getXMLSchemaCollections0() throws SQLException {
List result = new ArrayList<>();
return result;
}
@Override
protected List getUDTs0() throws SQLException {
List result = new ArrayList<>();
return result;
}
@Override
protected List getArrays0() throws SQLException {
List result = new ArrayList<>();
return result;
}
private Boolean is1_4_197;
private Boolean is1_4_198;
private Boolean is2_0_202;
boolean is1_4_197() {
// [#5874] The COLUMNS.COLUMN_TYPE column was introduced in H2 1.4.197 but removed again in H2 2.0.202
if (is1_4_197 == null)
is1_4_197 = configuredDialectIsNotFamilyAndSupports(asList(H2), () -> exists(COLUMNS.COLUMN_TYPE));
return is1_4_197;
}
boolean is1_4_198() {
// [#5874] The COLUMNS.IS_VISIBLE column was introduced in H2 1.4.198
if (is1_4_198 == null)
is1_4_198 = configuredDialectIsNotFamilyAndSupports(asList(H2), () -> exists(COLUMNS.IS_VISIBLE));
return is1_4_198;
}
boolean is2_0_202() {
// [#9609] The view was removed in H2 2.0: https://github.com/h2database/h2database/issues/2286
if (is2_0_202 == null)
is2_0_202 = configuredDialectIsNotFamilyAndSupports(asList(H2), () -> !exists(CONSTRAINTS));
return is2_0_202;
}
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy