com.pivotal.gemfirexd.internal.impl.sql.compile.QueryTreeNode Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of snappydata-store-core Show documentation
Show all versions of snappydata-store-core Show documentation
TIBCO ComputeDB store based off Pivotal GemFireXD
/*
Derby - Class com.pivotal.gemfirexd.internal.impl.sql.compile.QueryTreeNode
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to you under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
/*
* Changes for GemFireXD distributed data platform (some marked by "GemStone changes")
*
* Portions Copyright (c) 2010-2015 Pivotal Software, Inc. All rights reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License"); you
* may not use this file except in compliance with the License. You
* may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
* implied. See the License for the specific language governing
* permissions and limitations under the License. See accompanying
* LICENSE file.
*/
package com.pivotal.gemfirexd.internal.impl.sql.compile;
import java.sql.Types;
import java.util.List;
import com.pivotal.gemfirexd.internal.catalog.AliasInfo;
import com.pivotal.gemfirexd.internal.catalog.types.SynonymAliasInfo;
import com.pivotal.gemfirexd.internal.catalog.TypeDescriptor;
import com.pivotal.gemfirexd.internal.catalog.types.RowMultiSetImpl;
import com.pivotal.gemfirexd.internal.catalog.types.UserDefinedTypeIdImpl;
import com.pivotal.gemfirexd.internal.engine.distributed.metadata.QueryInfo;
import com.pivotal.gemfirexd.internal.engine.distributed.metadata.QueryInfoConstants;
import com.pivotal.gemfirexd.internal.engine.distributed.metadata.QueryInfoContext;
import com.pivotal.gemfirexd.internal.engine.distributed.metadata.SelectQueryInfo;
import com.pivotal.gemfirexd.internal.iapi.error.StandardException;
import com.pivotal.gemfirexd.internal.iapi.reference.ClassName;
import com.pivotal.gemfirexd.internal.iapi.reference.JDBC40Translation;
import com.pivotal.gemfirexd.internal.iapi.reference.SQLState;
import com.pivotal.gemfirexd.internal.iapi.services.classfile.VMOpcode;
import com.pivotal.gemfirexd.internal.iapi.services.compiler.MethodBuilder;
import com.pivotal.gemfirexd.internal.iapi.services.context.ContextManager;
import com.pivotal.gemfirexd.internal.iapi.services.loader.ClassFactory;
import com.pivotal.gemfirexd.internal.iapi.services.loader.ClassInspector;
import com.pivotal.gemfirexd.internal.iapi.services.sanity.SanityManager;
import com.pivotal.gemfirexd.internal.iapi.sql.StatementType;
import com.pivotal.gemfirexd.internal.iapi.sql.compile.C_NodeTypes;
import com.pivotal.gemfirexd.internal.iapi.sql.compile.CompilerContext;
import com.pivotal.gemfirexd.internal.iapi.sql.compile.NodeFactory;
import com.pivotal.gemfirexd.internal.iapi.sql.compile.Parser;
import com.pivotal.gemfirexd.internal.iapi.sql.compile.TypeCompiler;
import com.pivotal.gemfirexd.internal.iapi.sql.compile.Visitable;
import com.pivotal.gemfirexd.internal.iapi.sql.compile.Visitor;
import com.pivotal.gemfirexd.internal.iapi.sql.conn.LanguageConnectionContext;
import com.pivotal.gemfirexd.internal.iapi.sql.depend.DependencyManager;
import com.pivotal.gemfirexd.internal.iapi.sql.dictionary.AliasDescriptor;
import com.pivotal.gemfirexd.internal.iapi.sql.dictionary.DataDictionary;
import com.pivotal.gemfirexd.internal.iapi.sql.dictionary.SchemaDescriptor;
import com.pivotal.gemfirexd.internal.iapi.sql.dictionary.TableDescriptor;
import com.pivotal.gemfirexd.internal.iapi.sql.execute.ConstantAction;
import com.pivotal.gemfirexd.internal.iapi.sql.execute.ExecutionFactory;
import com.pivotal.gemfirexd.internal.iapi.types.DataTypeDescriptor;
import com.pivotal.gemfirexd.internal.iapi.types.DataValueDescriptor;
import com.pivotal.gemfirexd.internal.iapi.types.TypeId;
import com.pivotal.gemfirexd.internal.impl.sql.execute.GenericConstantActionFactory;
import com.pivotal.gemfirexd.internal.impl.sql.execute.GenericExecutionFactory;
import com.pivotal.gemfirexd.internal.shared.common.StoredFormatIds;
/**
* QueryTreeNode is the root class for all query tree nodes. All
* query tree nodes inherit from QueryTreeNode except for those that extend
* QueryTreeNodeVector.
*
*/
public abstract class QueryTreeNode implements Visitable
{
public static final int AUTOINCREMENT_START_INDEX = 0;
public static final int AUTOINCREMENT_INC_INDEX = 1;
public static final int AUTOINCREMENT_IS_AUTOINCREMENT_INDEX = 2;
//Parser uses this static field to make a note if the autoincrement column
//is participating in create or alter table.
public static final int AUTOINCREMENT_CREATE_MODIFY = 3;
// GemStone changes BEGIN
// used to send a warning in case autoinc start is specified explicitly
public static final int HAS_AUTOINCREMENT_INC = 4;
// for alter column SET GENERATED ALWAYS AS IDENTITY GemFireXD extension
public static final int SET_AUTOINCREMENT = 5;
public static final int AUTOINCREMENT_IS_GENERATED_BY_DEFAULT = 6;
// GemStone changes END
private int beginOffset = -1; // offset into SQL input of the substring
// which this query node encodes.
private int endOffset = -1;
private int nodeType;
private ContextManager cm;
// GemStone changes BEGIN
protected LanguageConnectionContext lcc;
/* (original code)
private LanguageConnectionContext lcc;
*/
/**
* Default Constructor
*/
public QueryTreeNode() {
super();
}
/**
* Copy Constructor
*
* For use in Non Collocated Join
* Copy everything used till preprocess phase (before optimize phase)
*/
public QueryTreeNode(QueryTreeNode other) {
// Super class is Object
super();
this.nodeType = other.nodeType;
this.cm = other.cm;
this.lcc = other.lcc;
}
// GemStone changes END
private GenericConstantActionFactory constantActionFactory;
/**
* In Derby SQL Standard Authorization, views, triggers and constraints
* execute with definer's privileges. Taking a specific eg of views
* user1
* create table t1 (c11 int);
* create view v1 as select * from user1.t1;
* grant select on v1 to user2;
* user2
* select * from user1.v1;
* Running with definer's privileges mean that since user2 has select
* privileges on view v1 owned by user1, then that is sufficient for user2
* to do a select from view v1. View v1 underneath might access some
* objects that user2 doesn't have privileges on, but that is not a problem
* since views execute with definer's privileges. In order to implement this
* behavior, when doing a select from view v1, we only want to check for
* select privilege on view v1. While processing the underlying query for
* view v1, we want to stop collecting the privilege requirements for the
* query underneath. Following flag, isPrivilegeCollectionRequired is used
* for this purpose. The flag will be true when we are the top level of view
* and then it is turned off while we process the query underlying the view
* v1.
*/
boolean isPrivilegeCollectionRequired = true;
/**
* Set the ContextManager for this node.
*
* @param cm The ContextManager.
*/
public void setContextManager(ContextManager cm)
{
this.cm = cm;
if (SanityManager.DEBUG)
{
SanityManager.ASSERT(cm != null,
"cm not expected to be null");
}
}
/**
* Get the current ContextManager.
*
* @return The current ContextManager.
*/
public final ContextManager getContextManager()
{
if (SanityManager.DEBUG) {
if (cm == null)
SanityManager.THROWASSERT("Null context manager in QueryTreeNode of type :" + this.getClass());
}
return cm;
}
/**
* Gets the NodeFactory for this database.
*
* @return the node factory for this database.
*
*/
public final NodeFactory getNodeFactory()
{
return getLanguageConnectionContext().getLanguageConnectionFactory().
getNodeFactory();
}
/**
* Gets the constant action factory for this database.
*
* @return the constant action factory.
*/
public final GenericConstantActionFactory getGenericConstantActionFactory()
{
if ( constantActionFactory == null )
{
GenericExecutionFactory execFactory = (GenericExecutionFactory) getExecutionFactory();
constantActionFactory = execFactory.getConstantActionFactory();
}
return constantActionFactory;
}
public final ExecutionFactory getExecutionFactory()
{
ExecutionFactory ef = getLanguageConnectionContext().getLanguageConnectionFactory().getExecutionFactory();
return ef;
}
/**
Get the ClassFactory to use with this database.
*/
protected final ClassFactory getClassFactory() {
return getLanguageConnectionContext().getLanguageConnectionFactory().
getClassFactory();
}
/**
* Gets the LanguageConnectionContext for this connection.
*
* @return the lcc for this connection
*
*/
// GemStone changes BEGIN
// made public
public final LanguageConnectionContext getLanguageConnectionContext()
/* (original code)
protected final LanguageConnectionContext getLanguageConnectionContext()
*/
// GemStone changes END
{
if (lcc == null)
{
lcc = (LanguageConnectionContext) getContextManager().
getContext(LanguageConnectionContext.CONTEXT_ID);
}
return lcc;
}
/**
* Gets the beginning offset of the SQL substring which this
* query node represents.
*
* @return The beginning offset of the SQL substring. -1 means unknown.
*
*/
public int getBeginOffset() { return beginOffset; }
/**
* Sets the beginning offset of the SQL substring which this
* query node represents.
*
* @param beginOffset The beginning offset of the SQL substring.
*
*/
public void setBeginOffset( int beginOffset )
{
this.beginOffset = beginOffset;
}
/**
* Gets the ending offset of the SQL substring which this
* query node represents.
*
* @return The ending offset of the SQL substring. -1 means unknown.
*
*/
public int getEndOffset() { return endOffset; }
/**
* Sets the ending offset of the SQL substring which this
* query node represents.
*
* @param endOffset The ending offset of the SQL substring.
*
*/
public void setEndOffset( int endOffset )
{
this.endOffset = endOffset;
}
/**
* Return header information for debug printing of this query
* tree node.
*
* @return Header information for debug printing of this query
* tree node.
*/
protected String nodeHeader()
{
if (SanityManager.DEBUG)
{
return "\n" + this.getClass().getName() + '@' +
Integer.toHexString(hashCode()) + "\n";
}
else
{
return "";
}
}
/**
* Format a node that has been converted to a String for printing
* as part of a tree. This method indents the String to the given
* depth by inserting tabs at the beginning of the string, and also
* after every newline.
*
* @param nodeString The node formatted as a String
* @param depth The depth to indent the given node
*
* @return The node String reformatted with tab indentation
*/
public static String formatNodeString(String nodeString, int depth)
{
if (SanityManager.DEBUG)
{
StringBuilder nodeStringBuilder = new StringBuilder(nodeString);
int pos;
char c;
char[] indent = new char[depth];
/*
** Form an array of tab characters for indentation.
*/
while (depth > 0)
{
indent[depth - 1] = '\t';
depth--;
}
/* Indent the beginning of the string */
nodeStringBuilder.insert(0, indent);
/*
** Look for newline characters, except for the last character.
** We don't want to indent after the last newline.
*/
for (pos = 0; pos < nodeStringBuilder.length() - 1; pos++)
{
c = nodeStringBuilder.charAt(pos);
if (c == '\n')
{
/* Indent again after each newline */
nodeStringBuilder.insert(pos + 1, indent);
}
}
return nodeStringBuilder.toString();
}
else
{
return "";
}
}
/**
* Print this tree for debugging purposes. This recurses through
* all the sub-nodes and prints them indented by their depth in
* the tree.
*/
public void treePrint()
{
if (SanityManager.DEBUG)
{
debugPrint(nodeHeader());
debugPrint(formatNodeString(this.toString(), 0));
printSubNodes(0);
debugFlush();
}
}
/**
* Print this tree for debugging purposes. This recurses through
* all the sub-nodes and prints them indented by their depth in
* the tree, starting with the given indentation.
*
* @param depth The depth of this node in the tree, thus,
* the amount to indent it when printing it.
*/
public void treePrint(int depth)
{
if (SanityManager.DEBUG)
{
debugPrint(formatNodeString(nodeHeader(), depth));
debugPrint(formatNodeString(this.toString(), depth));
printSubNodes(depth);
}
}
/**
* Print a String for debugging
*
* @param outputString The String to print
*/
public static void debugPrint(String outputString)
{
if (SanityManager.DEBUG) {
SanityManager.GET_DEBUG_STREAM().print(outputString);
}
}
/**
* Flush the debug stream out
*/
protected static void debugFlush()
{
if (SanityManager.DEBUG) {
SanityManager.GET_DEBUG_STREAM().flush();
}
}
/**
* Print the sub-nodes of this node.
*
* Each sub-class of QueryTreeNode is expected to provide its own
* printSubNodes() method. In each case, it calls super.printSubNodes(),
* passing along its depth, to get the sub-nodes of the super-class.
* Then it prints its own sub-nodes by calling treePrint() on each
* of its members that is a type of QueryTreeNode. In each case where
* it calls treePrint(), it should pass "depth + 1" to indicate that
* the sub-node should be indented one more level when printing.
* Also, it should call printLabel() to print the name of each sub-node
* before calling treePrint() on the sub-node, so that the reader of
* the printed tree can tell what the sub-node is.
*
* This printSubNodes() exists in here merely to act as a backstop.
* In other words, the calls to printSubNodes() move up the type
* hierarchy, and in this node the calls stop.
*
* I would have liked to put the call to super.printSubNodes() in
* this super-class, but Java resolves "super" statically, so it
* wouldn't get to the right super-class.
*
* @param depth The depth to indent the sub-nodes
*/
public void printSubNodes(int depth)
{
}
/**
* Format this node as a string
*
* Each sub-class of QueryTreeNode should implement its own toString()
* method. In each case, toString() should format the class members
* that are not sub-types of QueryTreeNode (printSubNodes() takes care
* of following the references to sub-nodes, and toString() takes care
* of all members that are not sub-nodes). Newlines should be used
* liberally - one good way to do this is to have a newline at the
* end of each formatted member. It's also a good idea to put the
* name of each member in front of the formatted value. For example,
* the code might look like:
*
* "memberName: " + memberName + "\n" + ...
*
* @return This node formatted as a String
*/
public String toString()
{
return "";
}
/**
* Print the given label at the given indentation depth.
*
* @param depth The depth of indentation to use when printing
* the label
* @param label The String to print
*/
public void printLabel(int depth, String label)
{
if (SanityManager.DEBUG)
{
debugPrint(formatNodeString("<" + depth + "> " + label, depth));
}
}
/**
* Return true if the node references SESSION schema tables (temporary or permanent)
*
* @return true if references SESSION schema tables, else false
*
* @exception StandardException Thrown on error
*/
public boolean referencesSessionSchema()
throws StandardException
{
return false;
}
/**
* Checks if the passed schema descriptor is for SESSION schema
*
* @return true if the passed schema descriptor is for SESSION schema
*
* @exception StandardException Thrown on error
*/
final boolean isSessionSchema(SchemaDescriptor sd)
{
// GemStone changes BEGIN
return sd != null ? isSessionSchema(sd.getSchemaName()) : false;
/* (original code)
return isSessionSchema(sd.getSchemaName());
*/
// GemStone changes END
}
/**
* Checks if the passed schema name is for SESSION schema
*
* @return true if the passed schema name is for SESSION schema
*
* @exception StandardException Thrown on error
*/
final boolean isSessionSchema(String schemaName)
{
return SchemaDescriptor.STD_DECLARED_GLOBAL_TEMPORARY_TABLES_SCHEMA_NAME.equals(schemaName);
}
/**
* Triggers, constraints and views get executed with their definer's
* privileges and they can exist in the system only if their definers'
* still have all the privileges to creeate them. Based on this, any
* time a trigger/view/constraint is executing, we do not need to waste
* time in checking if the definer still has the right set of privileges.
* At compile time, we wil make sure that we do not collect the privilege
* requirement for objects accessed with definer privileges by calling the
* following method.
*/
public void disablePrivilegeCollection()
{
isPrivilegeCollectionRequired = false;
}
/**
* Return true from this method means that we need to collect privilege
* requirement for this node. For following cases, this method will
* return true.
* 1)execute view - collect privilege to access view but do not collect
* privilege requirements for objects accessed by actual view uqery
* 2)execute select - collect privilege requirements for objects accessed
* by select statement
* 3)create view - collect privileges for select statement : the select
* statement for create view falls under 2) category above.
*
* @return true if need to collect privilege requirement for this node
*/
public boolean isPrivilegeCollectionRequired()
{
return(isPrivilegeCollectionRequired);
}
/**
* Get the optimizer's estimate of the number of rows returned or affected
* for an optimized QueryTree.
*
* For non-optimizable statements (for example, CREATE TABLE),
* return 0. For optimizable statements, this method will be
* over-ridden in the statement's root node (DMLStatementNode
* in all cases we know about so far).
*
* @return 0L
*/
public long getRowEstimate()
{
return 0L;
}
/**
* Do the code generation for this node. This is a place-holder
* method - it should be over-ridden in the sub-classes.
*
* @param acb The ActivationClassBuilder for the class being built
* @param mb The method for the generated code to go into
*
* @exception StandardException Thrown on error
*/
protected void generate(
ActivationClassBuilder acb,
MethodBuilder mb)
throws StandardException
{
throw StandardException.newException(SQLState.LANG_UNABLE_TO_GENERATE,
this.nodeHeader());
}
/**
* Parameter info is stored in the compiler context.
* Hide this from the callers.
*
*
* @return null
*
* @exception StandardException on error
*/
public DataTypeDescriptor[] getParameterTypes()
throws StandardException
{
return getCompilerContext().getParameterTypes();
}
//Gemstone changes Begin
public List getOriginalParameterTypeCompilers()
throws StandardException
{
return getCompilerContext().getOriginalParameterTypeCompilers();
}
//Gemstone changes END
/**
* This creates a class that will do the work that's constant
* across all Executions of a PreparedStatement. It's up to
* our subclasses to override this method if they need to compile
* constant actions into PreparedStatements.
*
* @exception StandardException Thrown on failure
*/
public ConstantAction makeConstantAction() throws StandardException
{
return null;
}
/**
* Set the node type for this node.
*
* @param nodeType The node type.
*/
public void setNodeType(int nodeType)
{
this.nodeType = nodeType;
}
// GemStone changes BEGIN
//protected int getNodeType()
public int getNodeType()
// GemStone changes END
{
return nodeType;
}
/**
* For final nodes, return whether or not
* the node represents the specified nodeType.
*
* @param nodeType The nodeType of interest.
*
* @return Whether or not
* the node represents the specified nodeType.
*/
protected boolean isInstanceOf(int nodeType)
{
return (this.nodeType == nodeType);
}
/**
* Get the DataDictionary
*
* @return The DataDictionary
*
*/
public final DataDictionary getDataDictionary()
{
return getLanguageConnectionContext().getDataDictionary();
}
public final DependencyManager getDependencyManager()
{
return getDataDictionary().getDependencyManager();
}
/**
* Get the CompilerContext
*
* @return The CompilerContext
*/
// GemStone changes BEGIN
public final CompilerContext getCompilerContext()
// GemStone changes END
{
return (CompilerContext) getContextManager().
getContext(CompilerContext.CONTEXT_ID);
}
/**
* Get the TypeCompiler associated with the given TypeId
*
* @param typeId The TypeId to get a TypeCompiler for
*
* @return The corresponding TypeCompiler
*
*/
protected final TypeCompiler getTypeCompiler(TypeId typeId)
{
return
getCompilerContext().getTypeCompilerFactory().getTypeCompiler(typeId);
}
/**
* Accept a visitor, and call v.visit()
* on child nodes as necessary.
*
* @param v the visitor
*
* @exception StandardException on error
*/
public Visitable accept(Visitor v)
throws StandardException
{
return v.visit(this);
}
/**
* Accept a visitor on all child nodes. All sub-classes that add fields
* that should be visited, should override this method and call
* {@code accept(v)} on all visitable fields, as well as
* {@code super.acceptChildren(v)} to make sure all visitable fields
* defined by the super-class are accepted too.
*
* @param v the visitor
* @throws StandardException on errors raised by the visitor
*/
void acceptChildren(Visitor v) throws StandardException {
// no children
}
/**
* Get the int value of a Property
*
* @param value Property value as a String
* @param key Key value of property
*
* @return The int value of the property
*
* @exception StandardException Thrown on failure
*/
protected int getIntProperty(String value, String key)
throws StandardException
{
int intVal = -1;
try
{
intVal = Integer.parseInt(value);
}
catch (NumberFormatException nfe)
{
throw StandardException.newException(SQLState.LANG_INVALID_NUMBER_FORMAT_FOR_OVERRIDE,
value, key);
}
return intVal;
}
/**
** Parse the a SQL statement from the body
* of another SQL statement. Pushes and pops a
* separate CompilerContext to perform the compilation.
*/
StatementNode parseStatement(String sql, boolean internalSQL) throws StandardException
{
/*
** Get a new compiler context, so the parsing of the text
** doesn't mess up anything in the current context
*/
LanguageConnectionContext lcc = getLanguageConnectionContext();
CompilerContext newCC = lcc.pushCompilerContext();
if (internalSQL)
newCC.setReliability(CompilerContext.INTERNAL_SQL_LEGAL);
// GemStone changes BEGIN
//[sb] disabling optimize literal for the underlying statement.
//newCC.switchOptimizeLiteral(false);
newCC.allowOptimizeLiteral(false);
// GemStone changes END
try
{
Parser p = newCC.getParser();
return p.parseStatement(sql);
}
finally
{
lcc.popCompilerContext(newCC);
}
}
// GemStone changes BEGIN
/**
* Return the type of statement, something from
* StatementType.
*
* @return the type of statement
*/
public int getStatementType()
{
return StatementType.UNKNOWN;
}
// GemStone changes END
public boolean foundString(String[] list, String search)
{
if (list == null)
{
return false;
}
for (int i = 0; i < list.length; i++)
{
if (list[i].equals(search))
{
return true;
}
}
return false;
}
/**
* Get a ConstantNode to represent a typed null value.
*
* @param type Type of the null node.
*
* @return A ConstantNode with the specified type, and a value of null
*
* @exception StandardException Thrown on error
*/
public ConstantNode getNullNode(DataTypeDescriptor type)
throws StandardException
{
int constantNodeType;
switch (type.getTypeId().getJDBCTypeId())
{
case Types.VARCHAR:
constantNodeType = C_NodeTypes.VARCHAR_CONSTANT_NODE;
break;
case Types.CHAR:
constantNodeType = C_NodeTypes.CHAR_CONSTANT_NODE;
break;
case Types.TINYINT:
constantNodeType = C_NodeTypes.TINYINT_CONSTANT_NODE;
break;
case Types.SMALLINT:
constantNodeType = C_NodeTypes.SMALLINT_CONSTANT_NODE;
break;
case Types.INTEGER:
constantNodeType = C_NodeTypes.INT_CONSTANT_NODE;
break;
case Types.BIGINT:
constantNodeType = C_NodeTypes.LONGINT_CONSTANT_NODE;
break;
case Types.REAL:
constantNodeType = C_NodeTypes.FLOAT_CONSTANT_NODE;
break;
case Types.DOUBLE:
constantNodeType = C_NodeTypes.DOUBLE_CONSTANT_NODE;
break;
case Types.NUMERIC:
case Types.DECIMAL:
constantNodeType = C_NodeTypes.DECIMAL_CONSTANT_NODE;
break;
case Types.DATE:
case Types.TIME:
case Types.TIMESTAMP:
constantNodeType = C_NodeTypes.USERTYPE_CONSTANT_NODE;
break;
case Types.BINARY:
constantNodeType = C_NodeTypes.BIT_CONSTANT_NODE;
break;
case Types.VARBINARY:
constantNodeType = C_NodeTypes.VARBIT_CONSTANT_NODE;
break;
case Types.LONGVARCHAR:
constantNodeType = C_NodeTypes.LONGVARCHAR_CONSTANT_NODE;
break;
case Types.CLOB:
constantNodeType = C_NodeTypes.CLOB_CONSTANT_NODE;
break;
case Types.LONGVARBINARY:
constantNodeType = C_NodeTypes.LONGVARBIT_CONSTANT_NODE;
break;
case Types.BLOB:
constantNodeType = C_NodeTypes.BLOB_CONSTANT_NODE;
break;
case JDBC40Translation.SQLXML:
constantNodeType = C_NodeTypes.XML_CONSTANT_NODE;
break;
case JDBC40Translation.JSON:
constantNodeType = C_NodeTypes.JSON_CONSTANT_NODE;
break;
case Types.BOOLEAN:
constantNodeType = C_NodeTypes.BOOLEAN_CONSTANT_NODE;
break;
default:
if (type.getTypeId().userType())
{
constantNodeType = C_NodeTypes.USERTYPE_CONSTANT_NODE;
}
else
{
if (SanityManager.DEBUG)
SanityManager.THROWASSERT( "Unknown type " +
type.getTypeId().getSQLTypeName() + " in getNullNode");
return null;
}
}
ConstantNode constantNode = (ConstantNode) getNodeFactory().getNode(
constantNodeType,
type.getTypeId(),
cm);
constantNode.setType(type.getNullabilityType(true));
return constantNode;
}
/**
* Translate a Default node into a default value, given a type descriptor.
*
* @param typeDescriptor A description of the required data type.
*
* @exception StandardException Thrown on error
*/
public DataValueDescriptor convertDefaultNode(DataTypeDescriptor typeDescriptor)
throws StandardException
{
/*
** Override in cases where node type
** can be converted to default value.
*/
return null;
}
/* Initializable methods */
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Single-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Two-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Three-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Four-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Five-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5,
Object arg6) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Six-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5,
Object arg6,
Object arg7) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Seven-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5,
Object arg6,
Object arg7,
Object arg8) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Eight-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5,
Object arg6,
Object arg7,
Object arg8,
Object arg9) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Nine-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5,
Object arg6,
Object arg7,
Object arg8,
Object arg9,
Object arg10) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Ten-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5,
Object arg6,
Object arg7,
Object arg8,
Object arg9,
Object arg10,
Object arg11) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Eleven-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5,
Object arg6,
Object arg7,
Object arg8,
Object arg9,
Object arg10,
Object arg11,
Object arg12) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Twelve-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5,
Object arg6,
Object arg7,
Object arg8,
Object arg9,
Object arg10,
Object arg11,
Object arg12,
Object arg13) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Thirteen-argument init() not implemented for " + getClass().getName());
}
}
/**
* Initialize a query tree node.
*
* @exception StandardException Thrown on error
*/
public void init(Object arg1,
Object arg2,
Object arg3,
Object arg4,
Object arg5,
Object arg6,
Object arg7,
Object arg8,
Object arg9,
Object arg10,
Object arg11,
Object arg12,
Object arg13,
Object arg14) throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("Fourteen-argument init() not implemented for " + getClass().getName());
}
}
public TableName makeTableName
(
String schemaName,
String flatName
)
throws StandardException
{
return (TableName) getNodeFactory().getNode
(
C_NodeTypes.TABLE_NAME,
schemaName,
flatName,
getContextManager()
);
}
public boolean isAtomic() throws StandardException
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT("isAtomic should not be called for this class: " + getClass().getName());
}
return false;
}
public Object getCursorInfo() throws StandardException
{
return null;
}
/**
* Get the descriptor for the named table within the given schema.
* If the schema parameter is NULL, it looks for the table in the
* current (default) schema. Table descriptors include object ids,
* object types (table, view, etc.)
* If the schema is SESSION, then before looking into the data dictionary
* for persistent tables, it first looks into LCC for temporary tables.
* If no temporary table tableName found for the SESSION schema, then it goes and
* looks through the data dictionary for persistent table
* We added getTableDescriptor here so that we can look for non data dictionary
* tables(ie temp tables) here. Any calls to getTableDescriptor in data dictionary
* should be only for persistent tables
*
* @param tableName The name of the table to get the descriptor for
* @param schema The descriptor for the schema the table lives in.
* If null, use the current (default) schema.
*
* @return The descriptor for the table, null if table does not
* exist.
*
* @exception StandardException Thrown on failure
*/
protected final TableDescriptor getTableDescriptor(String tableName,
SchemaDescriptor schema)
throws StandardException
{
TableDescriptor retval;
//Following if means we are dealing with SESSION schema.
if (isSessionSchema(schema))
{
//First we need to look in the list of temporary tables to see if this table is a temporary table.
retval = getLanguageConnectionContext().getTableDescriptorForDeclaredGlobalTempTable(tableName);
if (retval != null)
return retval; //this is a temporary table
}
//Following if means we are dealing with SESSION schema and we are dealing with in-memory schema (ie there is no physical SESSION schema)
//If following if is true, it means SESSION.table is not a declared table & it can't be physical SESSION.table
//because there is no physical SESSION schema
if (schema.getUUID() == null)
return null;
//it is not a temporary table, so go through the data dictionary to find the physical persistent table
TableDescriptor td = getDataDictionary().getTableDescriptor(tableName, schema,
this.getLanguageConnectionContext().getTransactionCompile());
if (td == null || td.isSynonymDescriptor())
return null;
return td;
}
/**
* Get the descriptor for the named schema. If the schemaName
* parameter is NULL, it gets the descriptor for the current
* compilation schema.
*
* QueryTreeNodes must obtain schemas using this method or the two argument
* version of it. This is to ensure that the correct default compliation schema
* is returned and to allow determination of if the statement being compiled
* depends on the current schema.
*
* Schema descriptors include authorization ids and schema ids.
* SQL92 allows a schema to specify a default character set - we will
* not support this. Will check default schema for a match
* before scanning a system table.
*
* @param schemaName The name of the schema we're interested in.
* If the name is NULL, get the descriptor for the
* current compilation schema.
*
* @return The descriptor for the schema.
*
* @exception StandardException Thrown on error
*/
//Gemstone changes BEGIN (add protected)
protected final SchemaDescriptor getSchemaDescriptor(String schemaName)
//Gemstone changes END
throws StandardException
{
//return getSchemaDescriptor(schemaName, schemaName != null);
return getSchemaDescriptor(schemaName, true);
}
/**
* Get the descriptor for the named schema. If the schemaName
* parameter is NULL, it gets the descriptor for the current
* compilation schema.
*
* QueryTreeNodes must obtain schemas using this method or the single argument
* version of it. This is to ensure that the correct default compliation schema
* is returned and to allow determination of if the statement being compiled
* depends on the current schema.
*
* @param schemaName The name of the schema we're interested in.
* If the name is NULL, get the descriptor for the current compilation schema.
* @param raiseError True to raise an error if the schema does not exist,
* false to return null if the schema does not exist.
* @return Valid SchemaDescriptor or null if raiseError is false and the
* schema does not exist.
* @throws StandardException Schema does not exist and raiseError is true.
*/
//Gemstone changes BEGIN (add protected)
protected final SchemaDescriptor getSchemaDescriptor(String schemaName, boolean raiseError)
//Gemstone changes END
throws StandardException
{
/*
** Check for a compilation context. Sometimes
** there is a special compilation context in
** place to recompile something that may have
** been compiled against a different schema than
** the current schema (e.g views):
**
** CREATE SCHEMA x
** CREATE TABLE t
** CREATE VIEW vt as SEELCT * FROM t
** SET SCHEMA app
** SELECT * FROM X.vt
**
** In the above view vt must be compiled against
** the X schema.
*/
SchemaDescriptor sd = null;
boolean isCurrent = false;
boolean isCompilation = false;
if (schemaName == null) {
CompilerContext cc = getCompilerContext();
sd = cc.getCompilationSchema();
if (sd == null) {
// Set the compilation schema to be the default,
// notes that this query has schema dependencies.
sd = getLanguageConnectionContext().getDefaultSchema();
isCurrent = true;
cc.setCompilationSchema(sd);
}
else
{
isCompilation = true;
}
schemaName = sd.getSchemaName();
}
DataDictionary dataDictionary = getDataDictionary();
SchemaDescriptor sdCatalog = dataDictionary.getSchemaDescriptor(schemaName,
getLanguageConnectionContext().getTransactionCompile(), raiseError);
if (isCurrent || isCompilation) {
//if we are dealing with a SESSION schema and it is not physically
//created yet, then it's uuid is going to be null. DERBY-1706
//Without the getUUID null check below, following will give NPE
//set schema session; -- session schema has not been created yet
//create table t1(c11 int);
if (sdCatalog != null && sdCatalog.getUUID() != null)
{
// different UUID for default (current) schema than in catalog,
// so reset default schema.
if (!sdCatalog.getUUID().equals(sd.getUUID()))
{
if (isCurrent)
getLanguageConnectionContext().setDefaultSchema(sdCatalog);
getCompilerContext().setCompilationSchema(sdCatalog);
}
}
else
{
// this schema does not exist, so ensure its UUID is null.
sd.setUUID(null);
sdCatalog = sd;
}
}
return sdCatalog;
}
/**
* Resolve table/view reference to a synonym. May have to follow a synonym chain.
*
* @param tabName to match for a synonym
*
* @return Synonym TableName if a match is found, NULL otherwise.
*
* @exception StandardException Thrown on error
*/
public TableName resolveTableToSynonym(TableName tabName) throws StandardException
{
DataDictionary dd = getDataDictionary();
String nextSynonymTable = tabName.getTableName();
String nextSynonymSchema = tabName.getSchemaName();
boolean found = false;
CompilerContext cc = getCompilerContext();
// Circular synonym references should have been detected at the DDL time, so
// the following loop shouldn't loop forever.
for (;;)
{
SchemaDescriptor nextSD = getSchemaDescriptor(nextSynonymSchema, false);
if (nextSD == null || nextSD.getUUID() == null)
break;
AliasDescriptor nextAD = dd.getAliasDescriptor(nextSD.getUUID().toString(),
nextSynonymTable, AliasInfo.ALIAS_NAME_SPACE_SYNONYM_AS_CHAR);
if (nextAD == null)
break;
/* Query is dependent on the AliasDescriptor */
cc.createDependency(nextAD);
found = true;
SynonymAliasInfo info = ((SynonymAliasInfo)nextAD.getAliasInfo());
nextSynonymTable = info.getSynonymTable();
nextSynonymSchema = info.getSynonymSchema();
}
if (!found)
return null;
TableName tableName = new TableName();
tableName.init(nextSynonymSchema, nextSynonymTable);
return tableName;
}
/**
* Verify that a java class exists, is accessible (public)
* and not a class representing a primitive type.
* @param javaClassName The name of the java class to resolve.
*
*
* @exception StandardException Thrown on error
*/
void verifyClassExist(String javaClassName)
throws StandardException
{
ClassInspector classInspector = getClassFactory().getClassInspector();
Throwable reason = null;
boolean foundMatch = false;
try {
foundMatch = classInspector.accessible(javaClassName);
} catch (ClassNotFoundException cnfe) {
reason = cnfe;
}
if (!foundMatch)
throw StandardException.newException(SQLState.LANG_TYPE_DOESNT_EXIST2, reason, javaClassName);
if (ClassInspector.primitiveType(javaClassName))
throw StandardException.newException(SQLState.LANG_TYPE_DOESNT_EXIST3, javaClassName);
}
/**
* set the Information gathered from the parent table that is
* required to peform a referential action on dependent table.
*/
public void setRefActionInfo(long fkIndexConglomId,
int[]fkColArray,
String parentResultSetId,
boolean dependentScan)
{
if (SanityManager.DEBUG)
{
SanityManager.THROWASSERT(
"setRefActionInfo() not expected to be called for " +
getClass().getName());
}
}
/**
Add an authorization check into the passed in method.
*/
// GemStone changes BEGIN
/*(original code) protected void generateAuthorizeCheck(ActivationClassBuilder acb, */
protected void generateAuthorizeCheck(ExpressionClassBuilder acb,
MethodBuilder mb,
int sqlOperation) {
// GemStone changes END
// add code to authorize statement execution.
acb.pushThisAsActivation(mb);
mb.callMethod(VMOpcode.INVOKEINTERFACE, null, "getLanguageConnectionContext",
ClassName.LanguageConnectionContext, 0);
mb.callMethod(VMOpcode.INVOKEINTERFACE, null, "getAuthorizer",
ClassName.Authorizer, 0);
acb.pushThisAsActivation(mb);
mb.push(sqlOperation);
mb.callMethod(VMOpcode.INVOKEINTERFACE, null, "authorize",
"void", 2);
}
/**
* Bind a UDT. This involves looking it up in the DataDictionary and filling
* in its class name.
*
* @param originalDTD A datatype: might be an unbound UDT and might not be
*
* @return The bound UDT if originalDTD was an unbound UDT; otherwise returns originalDTD.
*/
public DataTypeDescriptor bindUserType( DataTypeDescriptor originalDTD ) throws StandardException
{
// if the type is a table type, then we need to bind its user-typed columns
if ( originalDTD.getCatalogType().isRowMultiSet() ) { return bindRowMultiSet( originalDTD ); }
// nothing to do if this is not a user defined type
if ( !originalDTD.getTypeId().userType() ) { return originalDTD; }
UserDefinedTypeIdImpl userTypeID = (UserDefinedTypeIdImpl) originalDTD.getTypeId().getBaseTypeId();
// also nothing to do if the type has already been resolved
if ( userTypeID.isBound() ) { return originalDTD; }
// ok, we have an unbound UDT. lookup this type in the data dictionary
DataDictionary dd = getDataDictionary();
SchemaDescriptor typeSchema = getSchemaDescriptor( userTypeID.getSchemaName() );
char udtNameSpace = AliasInfo.ALIAS_NAME_SPACE_UDT_AS_CHAR;
String unqualifiedTypeName = userTypeID.getUnqualifiedName();
AliasDescriptor ad = dd.getAliasDescriptor( typeSchema.getUUID().toString(), unqualifiedTypeName, udtNameSpace );
if (ad == null)
{
throw StandardException.newException(SQLState.LANG_OBJECT_NOT_FOUND, ad.getAliasType(udtNameSpace), unqualifiedTypeName);
}
DataTypeDescriptor result = new DataTypeDescriptor
(
TypeId.getUserDefinedTypeId( typeSchema.getSchemaName(), unqualifiedTypeName, ad.getJavaClassName() ),
originalDTD.isNullable()
);
return result;
}
/**
* Bind the UDTs in a table type.
*
* @param originalDTD A datatype: might be an unbound UDT and might not be
*
* @return The bound table type if originalDTD was an unbound table type; otherwise returns originalDTD.
*/
public DataTypeDescriptor bindRowMultiSet( DataTypeDescriptor originalDTD ) throws StandardException
{
if ( !originalDTD.getCatalogType().isRowMultiSet() ) { return originalDTD; }
RowMultiSetImpl originalMultiSet = (RowMultiSetImpl) originalDTD.getTypeId().getBaseTypeId();
String[] columnNames = originalMultiSet.getColumnNames();
TypeDescriptor[] columnTypes = originalMultiSet.getTypes();
int columnCount = columnTypes.length;
for ( int i = 0; i < columnCount; i++ )
{
TypeDescriptor columnType = columnTypes[ i ];
if ( columnType.isUserDefinedType() )
{
DataTypeDescriptor newColumnDTD = DataTypeDescriptor.getType( columnType );
newColumnDTD = bindUserType( newColumnDTD );
TypeDescriptor newColumnType = newColumnDTD.getCatalogType();
// poke the bound type back into the multi set descriptor
columnTypes[ i ] = newColumnType;
}
}
return originalDTD;
}
/**
* Declare a dependency on a type. This is only used if the type is an ANSI UDT.
*
* @param dtd Type which may have a dependency declared on it.
*/
public void createTypeDependency( DataTypeDescriptor dtd ) throws StandardException
{
AliasDescriptor ad = getDataDictionary().getAliasDescriptorForUDT( null, dtd );
if ( ad != null ) { createTypeDependency( ad ); }
}
/**
* Declare a dependency on an ANSI UDT, identified by its AliasDescriptor,
* and check that you have privilege to use it.
*/
private void createTypeDependency( AliasDescriptor ad ) throws StandardException
{
if ( ad != null )
{
getCompilerContext().createDependency( ad );
}
}
// GemStone changes BEGIN
/**
* This method returns the appropriate QueryInfo Object obtained
* from the analysis of the optmized QueryTree.
* It needs to be appropriately overridden in the base classes
* @return QueryInfo Object
* @throws StandardException
*/
public QueryInfo computeQueryInfo(QueryInfoContext qic)throws StandardException
{
return QueryInfoConstants.DUMMY;
}
/**
* Create Query Info for InsertSubSelect case only
* As above but @ few places wanted to override computeQueryInfo,
* for example in @see ProjectRestrictNode.computeQueryInfo
* wanted to avoid setting this as root query info.
* @return QueryInfo Object
* @throws StandardException
*/
public QueryInfo computeSubSelectQueryInfo(QueryInfoContext qic)throws StandardException
{
return computeQueryInfo(qic);
}
//This is appropriately overriden in SingleChildResultSetNode
//and Join node classes
protected void optimizeForOffHeap( boolean shouldOptimize) {
}
// GemStone changes END
}