All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.datastax.driver.core.SchemaParser Maven / Gradle / Ivy

Go to download

A driver for Apache Cassandra 1.2+ that works exclusively with the Cassandra Query Language version 3 (CQL3) and Cassandra's binary protocol.

There is a newer version: 4.0.0
Show newest version
/*
 * Copyright DataStax, Inc.
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package com.datastax.driver.core;

import static com.datastax.driver.core.SchemaElement.AGGREGATE;
import static com.datastax.driver.core.SchemaElement.FUNCTION;
import static com.datastax.driver.core.SchemaElement.KEYSPACE;
import static com.datastax.driver.core.SchemaElement.TABLE;
import static com.datastax.driver.core.SchemaElement.TYPE;
import static com.datastax.driver.core.SchemaElement.VIEW;

import com.datastax.driver.core.exceptions.BusyConnectionException;
import com.datastax.driver.core.exceptions.ConnectionException;
import com.google.common.collect.Lists;
import com.google.common.collect.Maps;
import java.util.ArrayList;
import java.util.Collections;
import java.util.Comparator;
import java.util.HashMap;
import java.util.Iterator;
import java.util.LinkedHashMap;
import java.util.List;
import java.util.Map;
import java.util.concurrent.ExecutionException;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

abstract class SchemaParser {

  private static final Logger logger = LoggerFactory.getLogger(SchemaParser.class);

  private static final TypeCodec> LIST_OF_TEXT_CODEC =
      TypeCodec.list(TypeCodec.varchar());

  private static final SchemaParser V2_PARSER = new V2SchemaParser();
  private static final SchemaParser V3_PARSER = new V3SchemaParser();
  private static final SchemaParser V4_PARSER = new V4SchemaParser();

  static SchemaParser forVersion(VersionNumber cassandraVersion) {
    if (cassandraVersion.getMajor() >= 4) return V4_PARSER;
    if (cassandraVersion.getMajor() >= 3) return V3_PARSER;
    return V2_PARSER;
  }

  static SchemaParser forDseVersion(VersionNumber dseVersion) {
    if (dseVersion.getMajor() == 6 && dseVersion.getMinor() >= 8) return V4_PARSER;
    if (dseVersion.getMajor() >= 5) return V3_PARSER;
    return V2_PARSER;
  }

  abstract SystemRows fetchSystemRows(
      Cluster cluster,
      SchemaElement targetType,
      String targetKeyspace,
      String targetName,
      List targetSignature,
      Connection connection,
      VersionNumber cassandraVersion)
      throws ConnectionException, BusyConnectionException, ExecutionException, InterruptedException;

  abstract String tableNameColumn();

  void refresh(
      Cluster cluster,
      SchemaElement targetType,
      String targetKeyspace,
      String targetName,
      List targetSignature,
      Connection connection,
      VersionNumber cassandraVersion)
      throws ConnectionException, BusyConnectionException, ExecutionException,
          InterruptedException {

    SystemRows rows =
        fetchSystemRows(
            cluster,
            targetType,
            targetKeyspace,
            targetName,
            targetSignature,
            connection,
            cassandraVersion);

    Metadata metadata;
    try {
      metadata = cluster.getMetadata();
    } catch (IllegalStateException e) {
      logger.warn("Unable to refresh metadata, cluster has been closed");
      return;
    }
    metadata.lock.lock();
    try {
      if (targetType == null || targetType == KEYSPACE) {
        // building the whole schema or a keyspace
        assert rows.keyspaces != null;
        Map keyspaces = buildKeyspaces(rows, cassandraVersion, cluster);
        updateKeyspaces(metadata, metadata.keyspaces, keyspaces, targetKeyspace);
        // If we rebuild all from scratch or have an updated keyspace, rebuild the token map
        // since some replication on some keyspace may have changed
        metadata.rebuildTokenMap();
      } else {
        assert targetKeyspace != null;
        KeyspaceMetadata keyspace = metadata.keyspaces.get(targetKeyspace);
        // If we update a keyspace we don't know about, something went
        // wrong. Log an error and schedule a full schema rebuild.
        if (keyspace == null) {
          logger.info(
              String.format(
                  "Asked to rebuild %s %s.%s but I don't know keyspace %s",
                  targetType, targetKeyspace, targetName, targetKeyspace));
          metadata.cluster.submitSchemaRefresh(null, null, null, null);
        } else {
          switch (targetType) {
            case TABLE:
              if (rows.tables.containsKey(targetKeyspace)) {
                Map tables =
                    buildTables(
                        keyspace,
                        rows.tables.get(targetKeyspace),
                        rows.columns.get(targetKeyspace),
                        rows.indexes.get(targetKeyspace),
                        cassandraVersion,
                        cluster);
                updateTables(metadata, keyspace.tables, tables, targetName);
              }
              if (rows.views.containsKey(targetKeyspace)) {
                Map tables =
                    buildViews(
                        keyspace,
                        rows.views.get(targetKeyspace),
                        rows.columns.get(targetKeyspace),
                        cassandraVersion,
                        cluster);
                updateViews(metadata, keyspace.views, tables, targetName);
              }
              break;
            case TYPE:
              if (rows.udts.containsKey(targetKeyspace)) {
                Map userTypes =
                    buildUserTypes(
                        keyspace, rows.udts.get(targetKeyspace), cassandraVersion, cluster);
                updateUserTypes(metadata, keyspace.userTypes, userTypes, targetName);
              }
              break;
            case FUNCTION:
              if (rows.functions.containsKey(targetKeyspace)) {
                Map functions =
                    buildFunctions(
                        keyspace, rows.functions.get(targetKeyspace), cassandraVersion, cluster);
                updateFunctions(metadata, keyspace.functions, functions, targetName);
              }
              break;
            case AGGREGATE:
              if (rows.aggregates.containsKey(targetKeyspace)) {
                Map aggregates =
                    buildAggregates(
                        keyspace, rows.aggregates.get(targetKeyspace), cassandraVersion, cluster);
                updateAggregates(metadata, keyspace.aggregates, aggregates, targetName);
              }
              break;
          }
        }
      }
    } catch (RuntimeException e) {
      // Failure to parse the schema is definitively wrong so log a full-on error, but this won't
      // generally prevent queries to
      // work and this can happen when new Cassandra versions modify stuff in the schema and the
      // driver hasn't yet be modified.
      // So log, but let things go otherwise.
      logger.error(
          "Error parsing schema from Cassandra system tables: the schema in Cluster#getMetadata() will appear incomplete or stale",
          e);
    } finally {
      metadata.lock.unlock();
    }
  }

  private Map buildKeyspaces(
      SystemRows rows, VersionNumber cassandraVersion, Cluster cluster) {

    Map keyspaces = new LinkedHashMap();
    for (Row keyspaceRow : rows.keyspaces) {
      KeyspaceMetadata keyspace = KeyspaceMetadata.build(keyspaceRow, cassandraVersion);
      Map userTypes =
          buildUserTypes(keyspace, rows.udts.get(keyspace.getName()), cassandraVersion, cluster);
      for (UserType userType : userTypes.values()) {
        keyspace.add(userType);
      }
      Map tables =
          buildTables(
              keyspace,
              rows.tables.get(keyspace.getName()),
              rows.columns.get(keyspace.getName()),
              rows.indexes.get(keyspace.getName()),
              cassandraVersion,
              cluster);
      for (TableMetadata table : tables.values()) {
        keyspace.add(table);
      }
      Map functions =
          buildFunctions(
              keyspace, rows.functions.get(keyspace.getName()), cassandraVersion, cluster);
      for (FunctionMetadata function : functions.values()) {
        keyspace.add(function);
      }
      Map aggregates =
          buildAggregates(
              keyspace, rows.aggregates.get(keyspace.getName()), cassandraVersion, cluster);
      for (AggregateMetadata aggregate : aggregates.values()) {
        keyspace.add(aggregate);
      }
      Map views =
          buildViews(
              keyspace,
              rows.views.get(keyspace.getName()),
              rows.columns.get(keyspace.getName()),
              cassandraVersion,
              cluster);
      for (MaterializedViewMetadata view : views.values()) {
        keyspace.add(view);
      }
      keyspaces.put(keyspace.getName(), keyspace);
    }
    if (rows.virtualKeyspaces != null) {
      for (Row keyspaceRow : rows.virtualKeyspaces) {
        KeyspaceMetadata keyspace = KeyspaceMetadata.buildVirtual(keyspaceRow, cassandraVersion);
        Map tables =
            buildTables(
                keyspace,
                rows.virtualTables.get(keyspace.getName()),
                rows.virtualColumns.get(keyspace.getName()),
                Collections.>emptyMap(),
                cassandraVersion,
                cluster);
        for (TableMetadata table : tables.values()) {
          keyspace.add(table);
        }
        keyspaces.put(keyspace.getName(), keyspace);
      }
    }

    return keyspaces;
  }

  private Map buildTables(
      KeyspaceMetadata keyspace,
      List tableRows,
      Map> colsDefs,
      Map> indexDefs,
      VersionNumber cassandraVersion,
      Cluster cluster) {
    Map tables = new LinkedHashMap();
    if (tableRows != null) {
      for (Row tableDef : tableRows) {
        String cfName = tableDef.getString(tableNameColumn());
        try {
          Map cols = colsDefs == null ? null : colsDefs.get(cfName);
          if (cols == null || cols.isEmpty()) {
            if (cassandraVersion.getMajor() >= 2) {
              // In C* >= 2.0, we should never have no columns metadata because at the very least we
              // should
              // have the metadata corresponding to the default CQL metadata. So if we don't have
              // any columns,
              // that can only mean that the table got creating concurrently with our schema
              // queries, and the
              // query for columns metadata reached the node before the table was persisted while
              // the table
              // metadata one reached it afterwards. We could make the query to the column metadata
              // sequential
              // with the table metadata instead of in parallel, but it's probably not worth making
              // it slower
              // all the time to avoid this race since 1) it's very very uncommon and 2) we can just
              // ignore the
              // incomplete table here for now and it'll get updated next time with no particular
              // consequence
              // (if the table creation was concurrent with our querying, we'll get a notifciation
              // later and
              // will reupdate the schema for it anyway). See JAVA-320 for why we need this.
              continue;
            } else {
              // C* 1.2 don't persists default CQL metadata, so it's possible not to have columns
              // (for thirft
              // tables). But in that case TableMetadata.build() knows how to handle it.
              cols = Collections.emptyMap();
            }
          }
          List cfIndexes = (indexDefs == null) ? null : indexDefs.get(cfName);
          TableMetadata table =
              TableMetadata.build(
                  keyspace,
                  tableDef,
                  cols,
                  cfIndexes,
                  tableNameColumn(),
                  cassandraVersion,
                  cluster);
          tables.put(table.getName(), table);
        } catch (RuntimeException e) {
          // See #refresh for why we'd rather not propagate this further
          logger.error(
              String.format(
                  "Error parsing schema for table %s.%s: "
                      + "Cluster.getMetadata().getKeyspace(\"%s\").getTable(\"%s\") will be missing or incomplete",
                  keyspace.getName(), cfName, keyspace.getName(), cfName),
              e);
        }
      }
    }
    return tables;
  }

  private Map buildUserTypes(
      KeyspaceMetadata keyspace,
      List udtRows,
      VersionNumber cassandraVersion,
      Cluster cluster) {
    Map userTypes = new LinkedHashMap();
    if (udtRows != null) {
      for (Row udtRow : maybeSortUdts(udtRows, cluster, keyspace.getName())) {
        UserType type = UserType.build(keyspace, udtRow, cassandraVersion, cluster, userTypes);
        userTypes.put(type.getTypeName(), type);
      }
    }
    return userTypes;
  }

  // Some schema versions require parsing UDTs in a specific order
  protected List maybeSortUdts(List udtRows, Cluster cluster, String keyspace) {
    return udtRows;
  }

  private Map buildFunctions(
      KeyspaceMetadata keyspace,
      List functionRows,
      VersionNumber cassandraVersion,
      Cluster cluster) {
    Map functions = new LinkedHashMap();
    if (functionRows != null) {
      for (Row functionRow : functionRows) {
        FunctionMetadata function =
            FunctionMetadata.build(keyspace, functionRow, cassandraVersion, cluster);
        if (function != null) {
          String name =
              Metadata.fullFunctionName(function.getSimpleName(), function.getArguments().values());
          functions.put(name, function);
        }
      }
    }
    return functions;
  }

  private Map buildAggregates(
      KeyspaceMetadata keyspace,
      List aggregateRows,
      VersionNumber cassandraVersion,
      Cluster cluster) {
    Map aggregates = new LinkedHashMap();
    if (aggregateRows != null) {
      for (Row aggregateRow : aggregateRows) {
        AggregateMetadata aggregate =
            AggregateMetadata.build(keyspace, aggregateRow, cassandraVersion, cluster);
        if (aggregate != null) {
          String name =
              Metadata.fullFunctionName(aggregate.getSimpleName(), aggregate.getArgumentTypes());
          aggregates.put(name, aggregate);
        }
      }
    }
    return aggregates;
  }

  private Map buildViews(
      KeyspaceMetadata keyspace,
      List viewRows,
      Map> colsDefs,
      VersionNumber cassandraVersion,
      Cluster cluster) {
    Map views =
        new LinkedHashMap();
    if (viewRows != null) {
      for (Row viewRow : viewRows) {
        String viewName = viewRow.getString("view_name");
        try {
          Map cols = colsDefs.get(viewName);
          if (cols == null || cols.isEmpty())
            continue; // we probably raced, we will update the metadata next time

          MaterializedViewMetadata view =
              MaterializedViewMetadata.build(keyspace, viewRow, cols, cassandraVersion, cluster);
          if (view != null) views.put(view.getName(), view);
        } catch (RuntimeException e) {
          // See #refresh for why we'd rather not propagate this further
          logger.error(
              String.format(
                  "Error parsing schema for view %s.%s: "
                      + "Cluster.getMetadata().getKeyspace(\"%s\").getView(\"%s\") will be missing or incomplete",
                  keyspace.getName(), viewName, keyspace.getName(), viewName),
              e);
        }
      }
    }
    return views;
  }

  // Update oldKeyspaces with the changes contained in newKeyspaces.
  // This method also takes care of triggering the relevant events
  private void updateKeyspaces(
      Metadata metadata,
      Map oldKeyspaces,
      Map newKeyspaces,
      String keyspaceToRebuild) {
    Iterator it = oldKeyspaces.values().iterator();
    while (it.hasNext()) {
      KeyspaceMetadata oldKeyspace = it.next();
      String keyspaceName = oldKeyspace.getName();
      // If we're rebuilding only a single keyspace, we should only consider that one
      // because newKeyspaces will only contain that keyspace.
      if ((keyspaceToRebuild == null || keyspaceToRebuild.equals(keyspaceName))
          && !newKeyspaces.containsKey(keyspaceName)) {
        it.remove();
        metadata.triggerOnKeyspaceRemoved(oldKeyspace);
      }
    }
    for (KeyspaceMetadata newKeyspace : newKeyspaces.values()) {
      KeyspaceMetadata oldKeyspace = oldKeyspaces.put(newKeyspace.getName(), newKeyspace);
      if (oldKeyspace == null) {
        metadata.triggerOnKeyspaceAdded(newKeyspace);
      } else if (!oldKeyspace.equals(newKeyspace)) {
        metadata.triggerOnKeyspaceChanged(newKeyspace, oldKeyspace);
      }
      Map oldTables =
          oldKeyspace == null
              ? new HashMap()
              : new HashMap(oldKeyspace.tables);
      updateTables(metadata, oldTables, newKeyspace.tables, null);
      Map oldTypes =
          oldKeyspace == null
              ? new HashMap()
              : new HashMap(oldKeyspace.userTypes);
      updateUserTypes(metadata, oldTypes, newKeyspace.userTypes, null);
      Map oldFunctions =
          oldKeyspace == null
              ? new HashMap()
              : new HashMap(oldKeyspace.functions);
      updateFunctions(metadata, oldFunctions, newKeyspace.functions, null);
      Map oldAggregates =
          oldKeyspace == null
              ? new HashMap()
              : new HashMap(oldKeyspace.aggregates);
      updateAggregates(metadata, oldAggregates, newKeyspace.aggregates, null);
      Map oldViews =
          oldKeyspace == null
              ? new HashMap()
              : new HashMap(oldKeyspace.views);
      updateViews(metadata, oldViews, newKeyspace.views, null);
    }
  }

  private void updateTables(
      Metadata metadata,
      Map oldTables,
      Map newTables,
      String tableToRebuild) {
    Iterator it = oldTables.values().iterator();
    while (it.hasNext()) {
      TableMetadata oldTable = it.next();
      String tableName = oldTable.getName();
      // If we're rebuilding only a single table, we should only consider that one
      // because newTables will only contain that table.
      if ((tableToRebuild == null || tableToRebuild.equals(tableName))
          && !newTables.containsKey(tableName)) {
        it.remove();
        metadata.triggerOnTableRemoved(oldTable);
      }
    }
    for (TableMetadata newTable : newTables.values()) {
      TableMetadata oldTable = oldTables.put(newTable.getName(), newTable);
      if (oldTable == null) {
        metadata.triggerOnTableAdded(newTable);
      } else {
        // if we're updating a table only, we need to copy views from old table to the new table.
        if (tableToRebuild != null) {
          for (MaterializedViewMetadata view : oldTable.getViews()) {
            view.setBaseTable(newTable);
          }
        }
        if (!oldTable.equals(newTable)) {
          metadata.triggerOnTableChanged(newTable, oldTable);
        }
      }
    }
  }

  private void updateUserTypes(
      Metadata metadata,
      Map oldTypes,
      Map newTypes,
      String typeToRebuild) {
    Iterator it = oldTypes.values().iterator();
    while (it.hasNext()) {
      UserType oldType = it.next();
      String typeName = oldType.getTypeName();
      if ((typeToRebuild == null || typeToRebuild.equals(typeName))
          && !newTypes.containsKey(typeName)) {
        it.remove();
        metadata.triggerOnUserTypeRemoved(oldType);
      }
    }
    for (UserType newType : newTypes.values()) {
      UserType oldType = oldTypes.put(newType.getTypeName(), newType);
      if (oldType == null) {
        metadata.triggerOnUserTypeAdded(newType);
      } else if (!newType.equals(oldType)) {
        metadata.triggerOnUserTypeChanged(newType, oldType);
      }
    }
  }

  private void updateFunctions(
      Metadata metadata,
      Map oldFunctions,
      Map newFunctions,
      String functionToRebuild) {
    Iterator it = oldFunctions.values().iterator();
    while (it.hasNext()) {
      FunctionMetadata oldFunction = it.next();
      String oldFunctionName =
          Metadata.fullFunctionName(
              oldFunction.getSimpleName(), oldFunction.getArguments().values());
      if ((functionToRebuild == null || functionToRebuild.equals(oldFunctionName))
          && !newFunctions.containsKey(oldFunctionName)) {
        it.remove();
        metadata.triggerOnFunctionRemoved(oldFunction);
      }
    }
    for (FunctionMetadata newFunction : newFunctions.values()) {
      String newFunctionName =
          Metadata.fullFunctionName(
              newFunction.getSimpleName(), newFunction.getArguments().values());
      FunctionMetadata oldFunction = oldFunctions.put(newFunctionName, newFunction);
      if (oldFunction == null) {
        metadata.triggerOnFunctionAdded(newFunction);
      } else if (!newFunction.equals(oldFunction)) {
        metadata.triggerOnFunctionChanged(newFunction, oldFunction);
      }
    }
  }

  private void updateAggregates(
      Metadata metadata,
      Map oldAggregates,
      Map newAggregates,
      String aggregateToRebuild) {
    Iterator it = oldAggregates.values().iterator();
    while (it.hasNext()) {
      AggregateMetadata oldAggregate = it.next();
      String oldAggregateName =
          Metadata.fullFunctionName(oldAggregate.getSimpleName(), oldAggregate.getArgumentTypes());
      if ((aggregateToRebuild == null || aggregateToRebuild.equals(oldAggregateName))
          && !newAggregates.containsKey(oldAggregateName)) {
        it.remove();
        metadata.triggerOnAggregateRemoved(oldAggregate);
      }
    }
    for (AggregateMetadata newAggregate : newAggregates.values()) {
      String newAggregateName =
          Metadata.fullFunctionName(newAggregate.getSimpleName(), newAggregate.getArgumentTypes());
      AggregateMetadata oldAggregate = oldAggregates.put(newAggregateName, newAggregate);
      if (oldAggregate == null) {
        metadata.triggerOnAggregateAdded(newAggregate);
      } else if (!newAggregate.equals(oldAggregate)) {
        metadata.triggerOnAggregateChanged(newAggregate, oldAggregate);
      }
    }
  }

  private void updateViews(
      Metadata metadata,
      Map oldViews,
      Map newViews,
      String viewToRebuild) {
    Iterator it = oldViews.values().iterator();
    while (it.hasNext()) {
      MaterializedViewMetadata oldView = it.next();
      String aggregateName = oldView.getName();
      if ((viewToRebuild == null || viewToRebuild.equals(aggregateName))
          && !newViews.containsKey(aggregateName)) {
        it.remove();
        metadata.triggerOnMaterializedViewRemoved(oldView);
      }
    }
    for (MaterializedViewMetadata newView : newViews.values()) {
      MaterializedViewMetadata oldView = oldViews.put(newView.getName(), newView);
      if (oldView == null) {
        metadata.triggerOnMaterializedViewAdded(newView);
      } else if (!newView.equals(oldView)) {
        metadata.triggerOnMaterializedViewChanged(newView, oldView);
      }
    }
  }

  static Map> groupByKeyspace(ResultSet rs) {
    if (rs == null) return Collections.emptyMap();

    Map> result = new HashMap>();
    for (Row row : rs) {
      String ksName = row.getString(KeyspaceMetadata.KS_NAME);
      List l = result.get(ksName);
      if (l == null) {
        l = new ArrayList();
        result.put(ksName, l);
      }
      l.add(row);
    }
    return result;
  }

  static Map>> groupByKeyspaceAndCf(ResultSet rs, String tableName) {
    if (rs == null) return Collections.emptyMap();

    Map>> result = Maps.newHashMap();
    for (Row row : rs) {
      String ksName = row.getString(KeyspaceMetadata.KS_NAME);
      String cfName = row.getString(tableName);
      Map> rowsByCf = result.get(ksName);
      if (rowsByCf == null) {
        rowsByCf = Maps.newHashMap();
        result.put(ksName, rowsByCf);
      }
      List l = rowsByCf.get(cfName);
      if (l == null) {
        l = Lists.newArrayList();
        rowsByCf.put(cfName, l);
      }
      l.add(row);
    }
    return result;
  }

  static Map>> groupByKeyspaceAndCf(
      ResultSet rs, VersionNumber cassandraVersion, String tableName) {
    if (rs == null) return Collections.emptyMap();

    Map>> result =
        new HashMap>>();
    for (Row row : rs) {
      String ksName = row.getString(KeyspaceMetadata.KS_NAME);
      String cfName = row.getString(tableName);
      Map> colsByCf = result.get(ksName);
      if (colsByCf == null) {
        colsByCf = new HashMap>();
        result.put(ksName, colsByCf);
      }
      Map l = colsByCf.get(cfName);
      if (l == null) {
        l = new HashMap();
        colsByCf.put(cfName, l);
      }
      ColumnMetadata.Raw c = ColumnMetadata.Raw.fromRow(row, cassandraVersion);
      l.put(c.name, c);
    }
    return result;
  }

  private static ResultSetFuture queryAsync(
      String query, Connection connection, ProtocolVersion protocolVersion)
      throws ConnectionException, BusyConnectionException {
    DefaultResultSetFuture future =
        new DefaultResultSetFuture(null, protocolVersion, new Requests.Query(query));
    connection.write(future);
    return future;
  }

  private static ResultSet get(ResultSetFuture future)
      throws InterruptedException, ExecutionException {
    return (future == null) ? null : future.get();
  }

  /**
   * The rows from the system tables that we want to parse to metadata classes. The format of these
   * rows depends on the Cassandra version, but our parsing code knows how to handle the
   * differences.
   */
  private static class SystemRows {
    final ResultSet keyspaces;
    final Map> tables;
    final Map>> columns;
    final Map> udts;
    final Map> functions;
    final Map> aggregates;
    final Map> views;
    final Map>> indexes;
    final ResultSet virtualKeyspaces;
    final Map> virtualTables;
    final Map>> virtualColumns;

    public SystemRows(
        ResultSet keyspaces,
        Map> tables,
        Map>> columns,
        Map> udts,
        Map> functions,
        Map> aggregates,
        Map> views,
        Map>> indexes,
        ResultSet virtualKeyspaces,
        Map> virtualTables,
        Map>> virtualColumns) {
      this.keyspaces = keyspaces;
      this.tables = tables;
      this.columns = columns;
      this.udts = udts;
      this.functions = functions;
      this.aggregates = aggregates;
      this.views = views;
      this.indexes = indexes;
      this.virtualKeyspaces = virtualKeyspaces;
      this.virtualTables = virtualTables;
      this.virtualColumns = virtualColumns;
    }
  }

  private static class V2SchemaParser extends SchemaParser {

    private static final String SELECT_KEYSPACES = "SELECT * FROM system.schema_keyspaces";
    private static final String SELECT_COLUMN_FAMILIES =
        "SELECT * FROM system.schema_columnfamilies";
    private static final String SELECT_COLUMNS = "SELECT * FROM system.schema_columns";
    private static final String SELECT_USERTYPES = "SELECT * FROM system.schema_usertypes";
    private static final String SELECT_FUNCTIONS = "SELECT * FROM system.schema_functions";
    private static final String SELECT_AGGREGATES = "SELECT * FROM system.schema_aggregates";

    private static final String CF_NAME = "columnfamily_name";

    @Override
    SystemRows fetchSystemRows(
        Cluster cluster,
        SchemaElement targetType,
        String targetKeyspace,
        String targetName,
        List targetSignature,
        Connection connection,
        VersionNumber cassandraVersion)
        throws ConnectionException, BusyConnectionException, ExecutionException,
            InterruptedException {

      boolean isSchemaOrKeyspace = (targetType == null || targetType == KEYSPACE);

      String whereClause = "";
      if (targetType != null) {
        whereClause = " WHERE keyspace_name = '" + targetKeyspace + '\'';
        if (targetType == TABLE) whereClause += " AND columnfamily_name = '" + targetName + '\'';
        else if (targetType == TYPE) whereClause += " AND type_name = '" + targetName + '\'';
        else if (targetType == FUNCTION)
          whereClause +=
              " AND function_name = '"
                  + targetName
                  + "' AND signature = "
                  + LIST_OF_TEXT_CODEC.format(targetSignature);
        else if (targetType == AGGREGATE)
          whereClause +=
              " AND aggregate_name = '"
                  + targetName
                  + "' AND signature = "
                  + LIST_OF_TEXT_CODEC.format(targetSignature);
      }

      ResultSetFuture ksFuture = null,
          udtFuture = null,
          cfFuture = null,
          colsFuture = null,
          functionsFuture = null,
          aggregatesFuture = null;

      ProtocolVersion protocolVersion =
          cluster.getConfiguration().getProtocolOptions().getProtocolVersion();

      if (isSchemaOrKeyspace)
        ksFuture = queryAsync(SELECT_KEYSPACES + whereClause, connection, protocolVersion);

      if (isSchemaOrKeyspace && supportsUdts(cassandraVersion) || targetType == TYPE)
        udtFuture = queryAsync(SELECT_USERTYPES + whereClause, connection, protocolVersion);

      if (isSchemaOrKeyspace || targetType == TABLE) {
        cfFuture = queryAsync(SELECT_COLUMN_FAMILIES + whereClause, connection, protocolVersion);
        colsFuture = queryAsync(SELECT_COLUMNS + whereClause, connection, protocolVersion);
      }

      if ((isSchemaOrKeyspace && supportsUdfs(cassandraVersion) || targetType == FUNCTION))
        functionsFuture = queryAsync(SELECT_FUNCTIONS + whereClause, connection, protocolVersion);

      if (isSchemaOrKeyspace && supportsUdfs(cassandraVersion) || targetType == AGGREGATE)
        aggregatesFuture = queryAsync(SELECT_AGGREGATES + whereClause, connection, protocolVersion);

      return new SystemRows(
          get(ksFuture),
          groupByKeyspace(get(cfFuture)),
          groupByKeyspaceAndCf(get(colsFuture), cassandraVersion, CF_NAME),
          groupByKeyspace(get(udtFuture)),
          groupByKeyspace(get(functionsFuture)),
          groupByKeyspace(get(aggregatesFuture)),
          // No views nor separate indexes table in Cassandra 2:
          Collections.>emptyMap(),
          Collections.>>emptyMap(),
          null,
          Collections.>emptyMap(),
          Collections.>>emptyMap());
    }

    @Override
    String tableNameColumn() {
      return CF_NAME;
    }

    private boolean supportsUdts(VersionNumber cassandraVersion) {
      return cassandraVersion.getMajor() > 2
          || (cassandraVersion.getMajor() == 2 && cassandraVersion.getMinor() >= 1);
    }

    private boolean supportsUdfs(VersionNumber cassandraVersion) {
      return cassandraVersion.getMajor() > 2
          || (cassandraVersion.getMajor() == 2 && cassandraVersion.getMinor() >= 2);
    }
  }

  private static class V3SchemaParser extends SchemaParser {

    protected static final String SELECT_KEYSPACES = "SELECT * FROM system_schema.keyspaces";
    protected static final String SELECT_TABLES = "SELECT * FROM system_schema.tables";
    protected static final String SELECT_COLUMNS = "SELECT * FROM system_schema.columns";
    protected static final String SELECT_USERTYPES = "SELECT * FROM system_schema.types";
    protected static final String SELECT_FUNCTIONS = "SELECT * FROM system_schema.functions";
    protected static final String SELECT_AGGREGATES = "SELECT * FROM system_schema.aggregates";
    protected static final String SELECT_INDEXES = "SELECT * FROM system_schema.indexes";
    protected static final String SELECT_VIEWS = "SELECT * FROM system_schema.views";

    private static final String TABLE_NAME = "table_name";

    @Override
    SystemRows fetchSystemRows(
        Cluster cluster,
        SchemaElement targetType,
        String targetKeyspace,
        String targetName,
        List targetSignature,
        Connection connection,
        VersionNumber cassandraVersion)
        throws ConnectionException, BusyConnectionException, ExecutionException,
            InterruptedException {

      boolean isSchemaOrKeyspace = (targetType == null || targetType == KEYSPACE);

      ResultSetFuture ksFuture = null,
          udtFuture = null,
          cfFuture = null,
          colsFuture = null,
          functionsFuture = null,
          aggregatesFuture = null,
          indexesFuture = null,
          viewsFuture = null;

      ProtocolVersion protocolVersion =
          cluster.getConfiguration().getProtocolOptions().getProtocolVersion();

      if (isSchemaOrKeyspace)
        ksFuture =
            queryAsync(
                SELECT_KEYSPACES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);

      if (isSchemaOrKeyspace || targetType == TYPE)
        udtFuture =
            queryAsync(
                SELECT_USERTYPES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);

      if (isSchemaOrKeyspace || targetType == TABLE) {
        cfFuture =
            queryAsync(
                SELECT_TABLES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
        colsFuture =
            queryAsync(
                SELECT_COLUMNS
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
        indexesFuture =
            queryAsync(
                SELECT_INDEXES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
        viewsFuture =
            queryAsync(
                SELECT_VIEWS
                    + whereClause(
                        targetType == TABLE ? VIEW : targetType,
                        targetKeyspace,
                        targetName,
                        targetSignature),
                connection,
                protocolVersion);
      }

      if (isSchemaOrKeyspace || targetType == FUNCTION)
        functionsFuture =
            queryAsync(
                SELECT_FUNCTIONS
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);

      if (isSchemaOrKeyspace || targetType == AGGREGATE)
        aggregatesFuture =
            queryAsync(
                SELECT_AGGREGATES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);

      return new SystemRows(
          get(ksFuture),
          groupByKeyspace(get(cfFuture)),
          groupByKeyspaceAndCf(get(colsFuture), cassandraVersion, TABLE_NAME),
          groupByKeyspace(get(udtFuture)),
          groupByKeyspace(get(functionsFuture)),
          groupByKeyspace(get(aggregatesFuture)),
          groupByKeyspace(get(viewsFuture)),
          groupByKeyspaceAndCf(get(indexesFuture), TABLE_NAME),
          null,
          Collections.>emptyMap(),
          Collections.>>emptyMap());
    }

    @Override
    String tableNameColumn() {
      return TABLE_NAME;
    }

    protected String whereClause(
        SchemaElement targetType,
        String targetKeyspace,
        String targetName,
        List targetSignature) {
      String whereClause = "";
      if (targetType != null) {
        whereClause = " WHERE keyspace_name = '" + targetKeyspace + '\'';
        if (targetType == TABLE) whereClause += " AND table_name = '" + targetName + '\'';
        else if (targetType == VIEW) whereClause += " AND view_name = '" + targetName + '\'';
        else if (targetType == TYPE) whereClause += " AND type_name = '" + targetName + '\'';
        else if (targetType == FUNCTION)
          whereClause +=
              " AND function_name = '"
                  + targetName
                  + "' AND argument_types = "
                  + LIST_OF_TEXT_CODEC.format(targetSignature);
        else if (targetType == AGGREGATE)
          whereClause +=
              " AND aggregate_name = '"
                  + targetName
                  + "' AND argument_types = "
                  + LIST_OF_TEXT_CODEC.format(targetSignature);
      }
      return whereClause;
    }

    // Used by maybeSortUdts to sort at each dependency group alphabetically.
    private static final Comparator sortByTypeName =
        new Comparator() {
          @Override
          public int compare(Row o1, Row o2) {
            String type1 = o1.getString(UserType.TYPE_NAME);
            String type2 = o2.getString(UserType.TYPE_NAME);

            if (type1 == null && type2 == null) {
              return 0;
            } else if (type2 == null) {
              return 1;
            } else if (type1 == null) {
              return -1;
            } else {
              return type1.compareTo(type2);
            }
          }
        };

    @Override
    protected List maybeSortUdts(List udtRows, Cluster cluster, String keyspace) {
      if (udtRows.size() < 2) return udtRows;

      // For C* 3+, user-defined type resolution must be done in proper order
      // to guarantee that nested UDTs get resolved
      DirectedGraph graph = new DirectedGraph(sortByTypeName, udtRows);
      for (Row from : udtRows) {
        for (Row to : udtRows) {
          if (from != to && dependsOn(to, from, cluster, keyspace)) graph.addEdge(from, to);
        }
      }
      return graph.topologicalSort();
    }

    private boolean dependsOn(Row udt1, Row udt2, Cluster cluster, String keyspace) {
      List fieldTypes = udt1.getList(UserType.COLS_TYPES, String.class);
      String typeName = udt2.getString(UserType.TYPE_NAME);
      for (String fieldTypeStr : fieldTypes) {
        // use shallow user types since some definitions might not be known at this stage
        DataType fieldType =
            DataTypeCqlNameParser.parse(fieldTypeStr, cluster, keyspace, null, null, false, true);
        if (references(fieldType, typeName)) return true;
      }
      return false;
    }

    private boolean references(DataType dataType, String typeName) {
      if (dataType instanceof UserType.Shallow
          && ((UserType.Shallow) dataType).typeName.equals(typeName)) return true;
      for (DataType arg : dataType.getTypeArguments()) {
        if (references(arg, typeName)) return true;
      }
      if (dataType instanceof TupleType) {
        for (DataType arg : ((TupleType) dataType).getComponentTypes()) {
          if (references(arg, typeName)) return true;
        }
      }
      return false;
    }
  }

  private static class V4SchemaParser extends V3SchemaParser {

    private static final String SELECT_VIRTUAL_KEYSPACES =
        "SELECT * FROM system_virtual_schema.keyspaces";
    private static final String SELECT_VIRTUAL_TABLES =
        "SELECT * FROM system_virtual_schema.tables";
    private static final String SELECT_VIRTUAL_COLUMNS =
        "SELECT * FROM system_virtual_schema.columns";

    private static final String TABLE_NAME = "table_name";

    @Override
    SystemRows fetchSystemRows(
        Cluster cluster,
        SchemaElement targetType,
        String targetKeyspace,
        String targetName,
        List targetSignature,
        Connection connection,
        VersionNumber cassandraVersion)
        throws ConnectionException, BusyConnectionException, ExecutionException,
            InterruptedException {

      boolean isSchemaOrKeyspace = (targetType == null || targetType == KEYSPACE);

      ResultSetFuture ksFuture = null,
          udtFuture = null,
          cfFuture = null,
          colsFuture = null,
          functionsFuture = null,
          aggregatesFuture = null,
          indexesFuture = null,
          viewsFuture = null,
          virtualKeyspacesFuture = null,
          virtualTableFuture = null,
          virtualColumnsFuture = null;

      ProtocolVersion protocolVersion =
          cluster.getConfiguration().getProtocolOptions().getProtocolVersion();

      if (isSchemaOrKeyspace) {
        ksFuture =
            queryAsync(
                SELECT_KEYSPACES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
        virtualKeyspacesFuture =
            queryAsync(
                SELECT_VIRTUAL_KEYSPACES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
        virtualColumnsFuture =
            queryAsync(
                SELECT_VIRTUAL_COLUMNS
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
        virtualTableFuture =
            queryAsync(
                SELECT_VIRTUAL_TABLES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
      }

      if (isSchemaOrKeyspace || targetType == TYPE) {
        udtFuture =
            queryAsync(
                SELECT_USERTYPES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
      }

      if (isSchemaOrKeyspace || targetType == TABLE) {
        cfFuture =
            queryAsync(
                SELECT_TABLES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
        colsFuture =
            queryAsync(
                SELECT_COLUMNS
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
        indexesFuture =
            queryAsync(
                SELECT_INDEXES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
        viewsFuture =
            queryAsync(
                SELECT_VIEWS
                    + whereClause(
                        targetType == TABLE ? VIEW : targetType,
                        targetKeyspace,
                        targetName,
                        targetSignature),
                connection,
                protocolVersion);
      }

      if (isSchemaOrKeyspace || targetType == FUNCTION) {
        functionsFuture =
            queryAsync(
                SELECT_FUNCTIONS
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
      }

      if (isSchemaOrKeyspace || targetType == AGGREGATE) {
        aggregatesFuture =
            queryAsync(
                SELECT_AGGREGATES
                    + whereClause(targetType, targetKeyspace, targetName, targetSignature),
                connection,
                protocolVersion);
      }

      return new SystemRows(
          get(ksFuture),
          groupByKeyspace(get(cfFuture)),
          groupByKeyspaceAndCf(get(colsFuture), cassandraVersion, TABLE_NAME),
          groupByKeyspace(get(udtFuture)),
          groupByKeyspace(get(functionsFuture)),
          groupByKeyspace(get(aggregatesFuture)),
          groupByKeyspace(get(viewsFuture)),
          groupByKeyspaceAndCf(get(indexesFuture), TABLE_NAME),
          get(virtualKeyspacesFuture),
          groupByKeyspace(get(virtualTableFuture)),
          groupByKeyspaceAndCf(get(virtualColumnsFuture), cassandraVersion, TABLE_NAME));
    }
  }
}




© 2015 - 2024 Weber Informatics LLC | Privacy Policy