org.apache.jackrabbit.oak.plugins.document.SweepHelper Maven / Gradle / Ivy
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.jackrabbit.oak.plugins.document;
import java.util.Collections;
import java.util.Map;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.atomic.AtomicLong;
import java.util.concurrent.atomic.AtomicReference;
import org.apache.jackrabbit.oak.commons.TimeDurationFormatter;
import org.apache.jackrabbit.oak.plugins.document.util.Utils;
import static com.google.common.collect.Lists.newArrayList;
import static org.apache.jackrabbit.oak.plugins.document.Collection.JOURNAL;
import static org.apache.jackrabbit.oak.plugins.document.Collection.NODES;
import static org.apache.jackrabbit.oak.plugins.document.UpdateOp.Operation.Type.REMOVE_MAP_ENTRY;
import static org.apache.jackrabbit.oak.plugins.document.UpdateOp.Operation.Type.SET_MAP_ENTRY;
import static org.apache.jackrabbit.oak.plugins.document.util.Utils.getRootDocument;
/**
* Helper class to perform a revision sweep for a given clusterId.
*/
public final class SweepHelper {
private static final TimeDurationFormatter TDF = TimeDurationFormatter.forLogging();
public static void sweep(final DocumentStore store,
final RevisionContext context,
final MissingLastRevSeeker seeker) {
int clusterId = context.getClusterId();
NodeDocument rootDoc = getRootDocument(store);
Revision lastRev = rootDoc.getLastRev().get(clusterId);
if (lastRev == null) {
throw new IllegalArgumentException("root document does not" +
" have a lastRev entry for clusterId " + clusterId);
}
final AtomicReference sweepRev = new AtomicReference<>(
rootDoc.getSweepRevisions().getRevision(clusterId));
if (sweepRev.get() == null) {
sweepRev.set(new Revision(0, 0, clusterId));
}
final AtomicLong sweepUpdates = new AtomicLong();
final AtomicLong bcUpdates = new AtomicLong();
final AtomicLong revertUpdates = new AtomicLong();
long time = System.currentTimeMillis();
NodeDocumentSweeper sweeper = new NodeDocumentSweeper(context, true);
sweeper.sweep(seeker.getCandidates(sweepRev.get().getTimestamp()),
new NodeDocumentSweepListener() {
@Override
public void sweepUpdate(Map updates)
throws DocumentStoreException {
// create an invalidate entry
JournalEntry inv = JOURNAL.newDocument(store);
inv.modified(updates.keySet());
Revision r = context.newRevision().asBranchRevision();
UpdateOp invOp = inv.asUpdateOp(r);
// and reference it from a regular entry
JournalEntry entry = JOURNAL.newDocument(store);
entry.invalidate(Collections.singleton(r));
Revision jRev = context.newRevision();
UpdateOp jOp = entry.asUpdateOp(jRev);
if (!store.create(JOURNAL, newArrayList(invOp, jOp))) {
String msg = "Unable to create journal entries for " +
"document invalidation.";
throw new DocumentStoreException(msg);
}
sweepRev.set(Utils.max(sweepRev.get(), jRev));
// now that journal entry is in place, perform the actual
// updates on the documents
store.createOrUpdate(NODES, newArrayList(updates.values()));
sweepUpdates.incrementAndGet();
trackStats(updates.values());
System.out.println("Sweeper updated " + updates.keySet());
}
private void trackStats(Iterable ops) {
for (UpdateOp op : ops) {
boolean bcUpdate = false;
boolean revertUpdate = false;
for (UpdateOp.Operation o : op.getChanges().values()) {
if (o.type == SET_MAP_ENTRY) {
bcUpdate = true;
} else if (o.type == REMOVE_MAP_ENTRY) {
revertUpdate = true;
}
}
if (bcUpdate) {
bcUpdates.incrementAndGet();
}
if (revertUpdate) {
revertUpdates.incrementAndGet();
}
}
}
});
time = System.currentTimeMillis() - time;
// write back sweep revision if none was there or revisions were swept
if (rootDoc.getSweepRevisions().getRevision(clusterId) == null
|| sweepUpdates.get() > 0) {
// sweepRev must be at least the current lastRev
sweepRev.set(Utils.max(sweepRev.get(), lastRev));
// perform a conditional update on the last seen lastRev
// to ensure the cluster node is still inactive
UpdateOp op = new UpdateOp(rootDoc.getId(), false);
op.equals("_lastRev",
new Revision(0, 0, clusterId),
lastRev.toString());
NodeDocument.setSweepRevision(op, sweepRev.get());
NodeDocument.setLastRev(op, sweepRev.get());
if (store.findAndUpdate(NODES, op) != null) {
System.out.println("Updated sweep revision to " +
sweepRev.get() + ". Branch commit markers added to " +
bcUpdates.get() + " documents. Reverted uncommitted " +
"changes on " + revertUpdates.get() + " documents. " +
"(" + formatMillis(time) + ")");
} else {
System.out.println("Sweep revision updated failed. " +
"Is clusterId " + clusterId + " active again?");
}
} else {
System.out.println("Revision sweep not needed for " +
"clusterId " + clusterId);
}
}
private static String formatMillis(long millis) {
return TDF.format(millis, TimeUnit.MILLISECONDS);
}
}