bt.torrent.data.DefaultDataWorker Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of bt-core Show documentation
Show all versions of bt-core Show documentation
BitTorrent Client Library (Core)
The newest version!
/*
* Copyright (c) 2016—2021 Andrei Tomashpolskiy and individual contributors.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package bt.torrent.data;
import bt.data.ChunkDescriptor;
import bt.data.ChunkVerifier;
import bt.data.DataDescriptor;
import bt.metainfo.TorrentId;
import bt.net.Peer;
import bt.net.buffer.BufferedData;
import bt.runtime.Config;
import bt.service.IRuntimeLifecycleBinder;
import bt.torrent.TorrentRegistry;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.util.concurrent.CompletableFuture;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
import java.util.concurrent.ThreadFactory;
import java.util.concurrent.atomic.AtomicInteger;
public class DefaultDataWorker implements DataWorker {
private static final Logger LOGGER = LoggerFactory.getLogger(DefaultDataWorker.class);
private static final Exception QUEUE_FULL_EXCEPTION = new IllegalStateException("Queue is overloaded");
private final TorrentRegistry torrentRegistry;
private final ChunkVerifier verifier;
private final BlockCache blockCache;
private final ExecutorService executor;
private final int maxPendingTasks;
private final AtomicInteger pendingTasksCount;
public DefaultDataWorker(IRuntimeLifecycleBinder lifecycleBinder,
TorrentRegistry torrentRegistry,
ChunkVerifier verifier,
BlockCache blockCache,
Config config) {
this.torrentRegistry = torrentRegistry;
this.verifier = verifier;
this.blockCache = blockCache;
this.executor = Executors.newSingleThreadExecutor(new ThreadFactory() {
private AtomicInteger i = new AtomicInteger();
@Override
public Thread newThread(Runnable r) {
return new Thread(r, String.format("%d.bt.torrent.data.worker-%d", config.getAcceptorPort(), i.incrementAndGet()));
}
});
this.maxPendingTasks = config.getMaxIOQueueSize();
this.pendingTasksCount = new AtomicInteger();
lifecycleBinder.onShutdown("Shutdown data worker", this.executor::shutdownNow);
}
@Override
public CompletableFuture addBlockRequest(TorrentId torrentId, Peer peer, int pieceIndex, int offset, int length) {
DataDescriptor data = getDataDescriptor(torrentId);
if (!data.getBitfield().isVerified(pieceIndex)) {
if (LOGGER.isDebugEnabled()) {
LOGGER.debug("Rejecting request to read block because the piece is not verified yet:" +
" piece index {" + pieceIndex + "}, offset {" + offset + "}, length {" + length + "}, peer {" + peer + "}");
}
return CompletableFuture.completedFuture(BlockRead.rejected(peer, pieceIndex, offset, length));
} else if (!tryIncrementTaskCount()) {
LOGGER.warn("Rejecting request to read block because the queue is full:" +
" piece index {"+pieceIndex+"}, offset {"+offset+"}, length {"+length+"}, peer {"+peer+"}");
return CompletableFuture.completedFuture(BlockRead.exceptional(peer,
QUEUE_FULL_EXCEPTION, pieceIndex, offset, length));
}
return CompletableFuture.supplyAsync(() -> {
try {
BlockReader blockReader = blockCache.get(torrentId, pieceIndex, offset, length);
return BlockRead.ready(peer, pieceIndex, offset, length, blockReader);
} catch (Throwable e) {
LOGGER.error("Failed to perform request to read block:" +
" piece index {" + pieceIndex + "}, offset {" + offset + "}, length {" + length + "}, peer {" + peer + "}", e);
return BlockRead.exceptional(peer, e, pieceIndex, offset, length);
} finally {
pendingTasksCount.decrementAndGet();
}
}, executor);
}
@Override
public CompletableFuture addBlock(TorrentId torrentId, Peer peer, int pieceIndex, int offset, BufferedData buffer) {
if (!tryIncrementTaskCount()) {
LOGGER.warn("Can't accept write block request -- queue is full");
buffer.dispose();
return CompletableFuture.completedFuture(BlockWrite.exceptional(peer,
QUEUE_FULL_EXCEPTION, pieceIndex, offset, buffer.length()));
}
return CompletableFuture.supplyAsync(() -> {
try {
DataDescriptor data = getDataDescriptor(torrentId);
ChunkDescriptor chunk = data.getChunkDescriptors().get(pieceIndex);
if (chunk.isComplete()) {
if (LOGGER.isDebugEnabled()) {
LOGGER.debug("Rejecting request to write block because" +
" the chunk is already complete and verified (or awaiting verification): " +
"piece index {" + pieceIndex + "}, offset {" + offset + "}, length {" + buffer.length() + "}");
}
return BlockWrite.rejected(peer, pieceIndex, offset, buffer.length());
}
chunk.getData().getSubrange(offset).putBytes(buffer.buffer());
if (LOGGER.isTraceEnabled()) {
LOGGER.trace("Successfully processed block: " +
"piece index {" + pieceIndex + "}, offset {" + offset + "}, length {" + buffer.length() + "}");
}
CompletableFuture verificationFuture = null;
if (chunk.isComplete()) {
verificationFuture = CompletableFuture.supplyAsync(() -> {
boolean verified = verifier.verify(chunk);
if (verified) {
data.getBitfield().markLocalPieceVerified(pieceIndex);
} else {
// reset data
chunk.clear();
}
return verified;
}, executor);
}
return BlockWrite.complete(peer, pieceIndex, offset, buffer.length(), verificationFuture);
} catch (Throwable e) {
return BlockWrite.exceptional(peer, e, pieceIndex, offset, buffer.length());
} finally {
pendingTasksCount.decrementAndGet();
buffer.dispose();
}
}, executor);
}
private boolean tryIncrementTaskCount() {
int newCount = pendingTasksCount.updateAndGet(oldCount -> {
if (oldCount == maxPendingTasks) {
return oldCount;
} else {
return oldCount + 1;
}
});
return newCount < maxPendingTasks;
}
private DataDescriptor getDataDescriptor(TorrentId torrentId) {
return torrentRegistry.getDescriptor(torrentId).get().getDataDescriptor();
}
}