All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.camel.component.hawtdb.HawtDBFile Maven / Gradle / Ivy

There is a newer version: 2.25.4
Show newest version
/**
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 *
 *      http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.camel.component.hawtdb;

import java.io.File;
import java.io.IOException;

import org.apache.camel.Service;
import org.fusesource.hawtbuf.Buffer;
import org.fusesource.hawtbuf.codec.BufferCodec;
import org.fusesource.hawtbuf.codec.IntegerCodec;
import org.fusesource.hawtbuf.codec.StringCodec;
import org.fusesource.hawtdb.api.BTreeIndexFactory;
import org.fusesource.hawtdb.api.OptimisticUpdateException;
import org.fusesource.hawtdb.api.SortedIndex;
import org.fusesource.hawtdb.api.Transaction;
import org.fusesource.hawtdb.api.TxPageFile;
import org.fusesource.hawtdb.api.TxPageFileFactory;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

/**
 * Manages access to a shared HawtDB file.
 * 

* Will by default not sync writes which allows it to be faster. * You can force syncing by setting the sync option to true. */ public class HawtDBFile extends TxPageFileFactory implements Service { private static final Logger LOG = LoggerFactory.getLogger(HawtDBFile.class); // the root which contains an index with name -> page for the real indexes private static final BTreeIndexFactory ROOT_INDEXES_FACTORY = new BTreeIndexFactory<>(); // the real indexes where we store persisted data in buffers private static final BTreeIndexFactory INDEX_FACTORY = new BTreeIndexFactory<>(); private TxPageFile pageFile; static { ROOT_INDEXES_FACTORY.setKeyCodec(StringCodec.INSTANCE); ROOT_INDEXES_FACTORY.setValueCodec(IntegerCodec.INSTANCE); //ROOT_INDEXES_FACTORY.setDeferredEncoding(true); // TODO: use false due CAMEL-3826 until root cause is fixed in hawtdb ROOT_INDEXES_FACTORY.setDeferredEncoding(false); INDEX_FACTORY.setKeyCodec(BufferCodec.INSTANCE); INDEX_FACTORY.setValueCodec(BufferCodec.INSTANCE); //INDEX_FACTORY.setDeferredEncoding(true); // TODO: use false due CAMEL-3826 until root cause is fixed in hawtdb INDEX_FACTORY.setDeferredEncoding(false); } public HawtDBFile() { setSync(false); } public void start() { if (getFile() == null) { throw new IllegalArgumentException("A file must be configured"); } if (LOG.isDebugEnabled()) { LOG.debug("Starting HawtDB using file: {}", getFile()); } open(); pageFile = getTxPageFile(); execute(new Work() { public Boolean execute(Transaction tx) { if (!tx.allocator().isAllocated(0)) { // if we just created the file, first allocated page should be 0 ROOT_INDEXES_FACTORY.create(tx); LOG.info("Aggregation repository data store created using file: " + getFile()); } else { SortedIndex indexes = ROOT_INDEXES_FACTORY.open(tx); LOG.info("Aggregation repository data store loaded using file: " + getFile() + " containing " + indexes.size() + " repositories."); } return true; } @Override public String toString() { return "Allocation repository file: " + getFile(); } }, true); } public void stop() { File file = getFile(); LOG.debug("Stopping HawtDB using file: {}", file); try { close(); } catch (IOException e) { LOG.warn("Error closing HawtDB file " + file + ". This exception will be ignored.", e); } pageFile = null; } public T execute(Work work) { return execute(work, true); } public T execute(Work work, boolean rollbackOnOptimisticUpdateException) { LOG.trace("Executing work +++ start +++ {}", work); Transaction tx = pageFile.tx(); T answer = doExecute(work, tx, pageFile, rollbackOnOptimisticUpdateException); LOG.trace("Executing work +++ done +++ {}", work); return answer; } public SortedIndex getRepositoryIndex(Transaction tx, String name, boolean create) { SortedIndex answer = null; SortedIndex indexes = ROOT_INDEXES_FACTORY.open(tx); Integer location = indexes.get(name); if (create && location == null) { // create it.. SortedIndex created = INDEX_FACTORY.create(tx); int page = created.getIndexLocation(); // add it to indexes so we can find it the next time indexes.put(name, page); LOG.debug("Created new repository index with name {} at location {}", name, page); answer = created; } else if (location != null) { LOG.trace("Repository index with name {} at location {}", name, location); answer = INDEX_FACTORY.open(tx, location); } LOG.trace("Repository index with name {} -> {}", name, answer); return answer; } private static T doExecute(Work work, Transaction tx, TxPageFile page, boolean handleOptimisticLockingException) { T answer = null; boolean done = false; int attempt = 0; while (!done) { try { // only log at DEBUG level if we are retrying if (attempt > 0) { LOG.debug("Attempt {} to execute work {}", attempt, work); } attempt++; // execute and get answer answer = work.execute(tx); LOG.trace("TX is read only: {} for executed work: {}", tx.isReadOnly(), work); // commit work tx.commit(); // and flush so we ensure data is spooled to disk page.flush(); // and we are done done = true; } catch (OptimisticUpdateException e) { if (handleOptimisticLockingException) { // retry as we hit an optimistic update error LOG.warn("OptimisticUpdateException occurred at attempt " + attempt + " executing work " + work + ". Will do rollback and retry."); // no harm doing rollback before retry and no wait is needed tx.rollback(); } else { // we must rollback and rethrow tx.rollback(); throw e; } } catch (RuntimeException e) { LOG.warn("Error executing work " + work + ". Will do rollback.", e); tx.rollback(); throw e; } } return answer; } }





© 2015 - 2025 Weber Informatics LLC | Privacy Policy