All Downloads are FREE. Search and download functionalities are using the official Maven repository.

ch.qos.logback.ext.loggly.io.DiscardingRollingOutputStream Maven / Gradle / Ivy

The newest version!
/**
 * Copyright (C) 2014 The logback-extensions developers ([email protected])
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *         http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package ch.qos.logback.ext.loggly.io;

import java.io.ByteArrayOutputStream;
import java.io.IOException;
import java.io.OutputStream;
import java.util.concurrent.BlockingDeque;
import java.util.concurrent.ConcurrentLinkedQueue;
import java.util.concurrent.LinkedBlockingDeque;
import java.util.concurrent.atomic.AtomicInteger;
import java.util.concurrent.locks.ReentrantLock;

/**
 * 

* Capped in-memory {@linkplain OutputStream} composed of a chain of {@linkplain ByteArrayOutputStream} called 'buckets'. *

*

* Each 'bucket' is limited in size (see {@link #maxBucketSizeInBytes}) and the total size of the {@linkplain OutputStream} * is bounded thanks to a discarding policy. An external component is expected to consume the filled buckets thanks to * {@link #getFilledBuckets()}. *

*

* Implementation decisions: *

*
    *
  • Why in-memory without offload on disk: offload on disk was possible with Google Guava's * FileBackedOutputStream but had the drawback to introduce a dependency. Loggly batch appender use case * should be OK with a pure in-memory approach.
  • *
* * @author Cyrille Le Clerc */ public class DiscardingRollingOutputStream extends OutputStream { public static final String LINE_SEPARATOR = System.getProperty("line.separator"); private ByteArrayOutputStream currentBucket; private final ReentrantLock currentBucketLock = new ReentrantLock(); private final BlockingDeque filledBuckets; private final ConcurrentLinkedQueue recycledBucketPool; private long maxBucketSizeInBytes; private final AtomicInteger discardedBucketCount = new AtomicInteger(); /** * @param maxBucketSizeInBytes maximum byte size of each bucket * @param maxBucketCount maximum number of buckets */ public DiscardingRollingOutputStream(int maxBucketSizeInBytes, int maxBucketCount) { if (maxBucketCount < 2) { throw new IllegalArgumentException("'maxBucketCount' must be >1"); } this.maxBucketSizeInBytes = maxBucketSizeInBytes; this.filledBuckets = new LinkedBlockingDeque(maxBucketCount); this.recycledBucketPool = new ConcurrentLinkedQueue(); this.currentBucket = newBucket(); } @Override public void write(int b) throws IOException { currentBucketLock.lock(); try { currentBucket.write(b); rollCurrentBucketIfNeeded(); } finally { currentBucketLock.unlock(); } } @Override public void write(byte[] b) throws IOException { currentBucketLock.lock(); try { currentBucket.write(b); rollCurrentBucketIfNeeded(); } finally { currentBucketLock.unlock(); } } @Override public void write(byte[] b, int off, int len) throws IOException { currentBucketLock.lock(); try { currentBucket.write(b, off, len); rollCurrentBucketIfNeeded(); } finally { currentBucketLock.unlock(); } } @Override public void flush() throws IOException { currentBucketLock.lock(); try { currentBucket.flush(); } finally { currentBucketLock.unlock(); } } /** * Close all the underlying buckets (current bucket, filled buckets and buckets from the recycled buckets pool). */ @Override public void close() { // no-op as ByteArrayOutputStream#close() is no op } /** * Roll current bucket if size threshold has been reached. */ private void rollCurrentBucketIfNeeded() { if (currentBucket.size() < maxBucketSizeInBytes) { return; } rollCurrentBucket(); } /** * Roll current bucket if size threshold has been reached. */ public void rollCurrentBucketIfNotEmpty() { if (currentBucket.size() == 0) { return; } rollCurrentBucket(); } /** * Moves the current active bucket to the list of filled buckets and defines a new one. * * The new active bucket is reused from the {@link #recycledBucketPool} pool if one is available or recreated. */ public void rollCurrentBucket() { currentBucketLock.lock(); try { boolean offered = filledBuckets.offer(currentBucket); if (offered) { onBucketRoll(currentBucket); } else { onBucketDiscard(currentBucket); discardedBucketCount.incrementAndGet(); } currentBucket = newBucket(); } finally { currentBucketLock.unlock(); } } /** * Designed for extension. * * @param discardedBucket the discarded bucket */ protected void onBucketDiscard(ByteArrayOutputStream discardedBucket) { } /** * The rolled bucket. Designed for extension. * * @param rolledBucket the discarded bucket */ protected void onBucketRoll(ByteArrayOutputStream rolledBucket) { } /** * Get a new bucket from the {@link #recycledBucketPool} or instantiate a new one if none available * in the free bucket pool. * * @return the bucket ready to use */ protected ByteArrayOutputStream newBucket() { ByteArrayOutputStream bucket = recycledBucketPool.poll(); if (bucket == null) { bucket = new ByteArrayOutputStream(); } return bucket; } /** * Returns the given bucket to the pool of free buckets. * * @param bucket the bucket to recycle */ public void recycleBucket(ByteArrayOutputStream bucket) { bucket.reset(); recycledBucketPool.offer(bucket); } /** * Return the filled buckets */ public BlockingDeque getFilledBuckets() { return filledBuckets; } /** * @return Number of discarded buckets. Monitoring oriented metric. */ public int getDiscardedBucketCount() { return discardedBucketCount.get(); } public long getCurrentOutputStreamSize() { long sizeInBytes = 0; for (ByteArrayOutputStream bucket : filledBuckets) { sizeInBytes += bucket.size(); } sizeInBytes += currentBucket.size(); return sizeInBytes; } @Override public String toString() { return "DiscardingRollingOutputStream{" + "currentBucket.bytesWritten=" + currentBucket.size() + ", filledBuckets.size=" + filledBuckets.size() + ", discardedBucketCount=" + discardedBucketCount + ", recycledBucketPool.size=" + recycledBucketPool.size() + '}'; } }




© 2015 - 2025 Weber Informatics LLC | Privacy Policy