All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.cayenne.access.DataPort Maven / Gradle / Ivy

There is a newer version: 2.0.4
Show newest version
/*****************************************************************
 *   Licensed to the Apache Software Foundation (ASF) under one
 *  or more contributor license agreements.  See the NOTICE file
 *  distributed with this work for additional information
 *  regarding copyright ownership.  The ASF licenses this file
 *  to you under the Apache License, Version 2.0 (the
 *  "License"); you may not use this file except in compliance
 *  with the License.  You may obtain a copy of the License at
 *
 *    http://www.apache.org/licenses/LICENSE-2.0
 *
 *  Unless required by applicable law or agreed to in writing,
 *  software distributed under the License is distributed on an
 *  "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 *  KIND, either express or implied.  See the License for the
 *  specific language governing permissions and limitations
 *  under the License.
 ****************************************************************/

package org.apache.cayenne.access;

import java.util.ArrayList;
import java.util.Collection;
import java.util.Collections;
import java.util.Iterator;
import java.util.List;
import java.util.Map;

import org.apache.cayenne.CayenneException;
import org.apache.cayenne.access.util.IteratedSelectObserver;
import org.apache.cayenne.map.DbEntity;
import org.apache.cayenne.map.DerivedDbEntity;
import org.apache.cayenne.query.InsertBatchQuery;
import org.apache.cayenne.query.Query;
import org.apache.cayenne.query.SQLTemplate;
import org.apache.cayenne.query.SelectQuery;

/**
 * An engine to port data between two DataNodes. These nodes can potentially connect to
 * databases from different vendors. The only assumption is that all of the DbEntities
 * (tables) being ported are present in both source and destination databases and are
 * adequately described by Cayenne mapping.
 * 

* DataPort implements a Cayenne-based algorithm to read data from source DataNode and * write to destination DataNode. It uses DataPortDelegate interface to externalize * various things, such as determining what entities to port (include/exclude from port * based on some criteria), logging the progress of port operation, qualifying the * queries, etc. *

* * @since 1.2: Prior to 1.2 DataPort classes were a part of cayenne-examples package. * @author Andrei Adamchik */ public class DataPort { public static final int INSERT_BATCH_SIZE = 1000; protected DataNode sourceNode; protected DataNode destinationNode; protected Collection entities; protected boolean cleaningDestination; protected DataPortDelegate delegate; protected int insertBatchSize; public DataPort() { this.insertBatchSize = INSERT_BATCH_SIZE; } /** * Creates a new DataPort instance, setting its delegate. */ public DataPort(DataPortDelegate delegate) { this.delegate = delegate; } /** * Runs DataPort. The instance must be fully configured by the time this method is * invoked, having its delegate, source and destinatio nodes, and a list of entities * set up. */ public void execute() throws CayenneException { // sanity check if (sourceNode == null) { throw new CayenneException("Can't port data, source node is null."); } if (destinationNode == null) { throw new CayenneException("Can't port data, destination node is null."); } // the simple equality check may actually detect problems with misconfigred nodes // it is not as dumb as it may look at first if (sourceNode == destinationNode) { throw new CayenneException( "Can't port data, source and target nodes are the same."); } if (entities == null || entities.isEmpty()) { return; } // sort entities for insertion List sorted = new ArrayList(entities); destinationNode.getEntitySorter().sortDbEntities(sorted, false); if (cleaningDestination) { // reverse insertion order for deletion List entitiesInDeleteOrder = new ArrayList(sorted.size()); entitiesInDeleteOrder.addAll(sorted); Collections.reverse(entitiesInDeleteOrder); processDelete(entitiesInDeleteOrder); } processInsert(sorted); } /** * Cleans up destination tables data. */ protected void processDelete(List entities) { // Allow delegate to modify the list of entities // any way it wants. For instance delegate may filter // or sort the list (though it doesn't have to, and can simply // pass through the original list). if (delegate != null) { entities = delegate.willCleanData(this, entities); } if (entities == null || entities.isEmpty()) { return; } // Using QueryResult as observer for the data cleanup. // This allows to collect query statistics and pass it to the delegate. QueryResult observer = new QueryResult(); // Delete data from entities one by one Iterator it = entities.iterator(); while (it.hasNext()) { DbEntity entity = (DbEntity) it.next(); // skip derived DbEntities. Should we consult delegate ? // Using derived entities may allow things like materialized views.... if (entity instanceof DerivedDbEntity) { continue; } Query query = new SQLTemplate(entity, "DELETE FROM " + entity.getFullyQualifiedName()); // notify delegate that delete is about to happen if (delegate != null) { query = delegate.willCleanData(this, entity, query); } // perform delete query observer.clear(); destinationNode.performQueries(Collections.singletonList(query), observer); // notify delegate that delete just happened if (delegate != null) { // observer will store query statistics int count = observer.getFirstUpdateCount(query); delegate.didCleanData(this, entity, count); } } } /** * Reads source data from source, saving it to destination. */ protected void processInsert(List entities) throws CayenneException { // Allow delegate to modify the list of entities // any way it wants. For instance delegate may filter // or sort the list (though it doesn't have to, and can simply // pass through the original list). if (delegate != null) { entities = delegate.willCleanData(this, entities); } if (entities == null || entities.isEmpty()) { return; } // Create an observer for to get the iterated result // instead of getting each table as a list IteratedSelectObserver observer = new IteratedSelectObserver(); // Using QueryResult as observer for the data insert. // This allows to collect query statistics and pass it to the delegate. QueryResult insertObserver = new QueryResult(); // process ordered list of entities one by one Iterator it = entities.iterator(); while (it.hasNext()) { insertObserver.clear(); DbEntity entity = (DbEntity) it.next(); // skip derived DbEntities... if (entity instanceof DerivedDbEntity) { continue; } SelectQuery select = new SelectQuery(entity); select.setFetchingDataRows(true); // delegate is allowed to substitute query Query query = (delegate != null) ? delegate.willPortEntity( this, entity, select) : select; sourceNode.performQueries(Collections.singletonList(query), observer); ResultIterator result = observer.getResultIterator(); InsertBatchQuery insert = new InsertBatchQuery(entity, INSERT_BATCH_SIZE); try { // Split insertions into the same table into batches. // This will allow to process tables of arbitrary size // and not run out of memory. int currentRow = 0; // even if we don't use intermediate batch commits, we still need to // estimate batch insert size int batchSize = insertBatchSize > 0 ? insertBatchSize : INSERT_BATCH_SIZE; while (result.hasNextRow()) { if (insertBatchSize > 0 && currentRow > 0 && currentRow % insertBatchSize == 0) { // end of the batch detected... commit and start a new insert // query destinationNode.performQueries( Collections.singletonList(insert), insertObserver); insert = new InsertBatchQuery(entity, batchSize); insertObserver.clear(); } currentRow++; Map nextRow = result.nextDataRow(); insert.add(nextRow); } // commit remaining batch if needed if (insert.size() > 0) { destinationNode.performQueries( Collections.singletonList(insert), insertObserver); } if (delegate != null) { delegate.didPortEntity(this, entity, currentRow); } } finally { try { // don't forget to close ResultIterator result.close(); } catch (CayenneException ex) { } } } } public Collection getEntities() { return entities; } public DataNode getSourceNode() { return sourceNode; } public DataNode getDestinationNode() { return destinationNode; } /** * Sets the initial list of entities to process. This list can be later modified by * the delegate. */ public void setEntities(Collection entities) { this.entities = entities; } /** * Sets the DataNode serving as a source of the ported data. */ public void setSourceNode(DataNode sourceNode) { this.sourceNode = sourceNode; } /** * Sets the DataNode serving as a destination of the ported data. */ public void setDestinationNode(DataNode destinationNode) { this.destinationNode = destinationNode; } /** * Returns previously initialized DataPortDelegate object. */ public DataPortDelegate getDelegate() { return delegate; } public void setDelegate(DataPortDelegate delegate) { this.delegate = delegate; } /** * Returns true if a DataPort was configured to delete all data from the destination * tables. */ public boolean isCleaningDestination() { return cleaningDestination; } /** * Defines whether DataPort should delete all data from destination tables before * doing the port. */ public void setCleaningDestination(boolean cleaningDestination) { this.cleaningDestination = cleaningDestination; } public int getInsertBatchSize() { return insertBatchSize; } /** * Sets a parameter used for tuning insert batches. If set to a value greater than * zero, DataPort will commit every N rows. If set to value less or equal to zero, * DataPort will commit only once at the end of the insert. */ public void setInsertBatchSize(int insertBatchSize) { this.insertBatchSize = insertBatchSize; } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy