All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.iceberg.spark.RemoveCarryoverIterator Maven / Gradle / Ivy

There is a newer version: 1.7.0
Show newest version
/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *   http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing,
 * software distributed under the License is distributed on an
 * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 * KIND, either express or implied.  See the License for the
 * specific language governing permissions and limitations
 * under the License.
 */
package org.apache.iceberg.spark;

import java.util.Iterator;
import java.util.Set;
import org.apache.iceberg.relocated.com.google.common.collect.Sets;
import org.apache.spark.sql.Row;
import org.apache.spark.sql.types.StructType;

/**
 * An iterator that removes the carry-over rows from changelog tables within a single Spark task. It
 * assumes that rows are partitioned by identifier(or all) columns, and it is sorted by both
 * identifier(or all) columns and change type.
 *
 * 

Carry-over rows are the result of a removal and insertion of the same row within an operation * because of the copy-on-write mechanism. For example, given a file which contains row1 (id=1, * data='a') and row2 (id=2, data='b'). A copy-on-write delete of row2 would require erasing this * file and preserving row1 in a new file. The change-log table would report this as follows, * despite it not being an actual change to the table. * *

    *
  • (id=1, data='a', op='DELETE') *
  • (id=1, data='a', op='INSERT') *
  • (id=2, data='b', op='DELETE') *
* * The iterator finds the carry-over rows and removes them from the result. For example, the above * rows will be converted to: * *
    *
  • (id=2, data='b', op='DELETE') *
*/ class RemoveCarryoverIterator extends ChangelogIterator { private final int[] indicesToIdentifySameRow; private Row cachedDeletedRow = null; private long deletedRowCount = 0; private Row cachedNextRecord = null; RemoveCarryoverIterator(Iterator rowIterator, StructType rowType) { super(rowIterator, rowType); this.indicesToIdentifySameRow = generateIndicesToIdentifySameRow(); } @Override public boolean hasNext() { if (hasCachedDeleteRow() || cachedNextRecord != null) { return true; } return rowIterator().hasNext(); } @Override public Row next() { Row currentRow; if (returnCachedDeleteRow()) { // Non-carryover delete rows found. One or more identical delete rows were seen followed by a // non-identical row. This means none of the delete rows were carry over rows. Emit one // delete row and decrease the amount of delete rows seen. deletedRowCount--; currentRow = cachedDeletedRow; if (deletedRowCount == 0) { cachedDeletedRow = null; } return currentRow; } else if (cachedNextRecord != null) { currentRow = cachedNextRecord; cachedNextRecord = null; } else { currentRow = rowIterator().next(); } // If the current row is a delete row, drain all identical delete rows if (changeType(currentRow).equals(DELETE) && rowIterator().hasNext()) { cachedDeletedRow = currentRow; deletedRowCount = 1; Row nextRow = rowIterator().next(); // drain all identical delete rows when there is at least one cached delete row and the next // row is the same record while (nextRow != null && cachedDeletedRow != null && isSameRecord(cachedDeletedRow, nextRow, indicesToIdentifySameRow)) { if (changeType(nextRow).equals(INSERT)) { deletedRowCount--; if (deletedRowCount == 0) { cachedDeletedRow = null; } } else { deletedRowCount++; } if (rowIterator().hasNext()) { nextRow = rowIterator().next(); } else { nextRow = null; } } cachedNextRecord = nextRow; return null; } else { // either there is no cached delete row or the current row is not a delete row return currentRow; } } /** * The iterator returns a cached delete row if there are delete rows cached and the next row is * not the same record or there is no next row. */ private boolean returnCachedDeleteRow() { return hitBoundary() && hasCachedDeleteRow(); } private boolean hitBoundary() { return !rowIterator().hasNext() || cachedNextRecord != null; } private boolean hasCachedDeleteRow() { return cachedDeletedRow != null; } private int[] generateIndicesToIdentifySameRow() { Set metadataColumnIndices = Sets.newHashSet(changeTypeIndex()); return generateIndicesToIdentifySameRow(rowType().size(), metadataColumnIndices); } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy