All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.jgroups.protocols.UNICAST3 Maven / Gradle / Ivy

Go to download

This artifact provides a single jar that contains all classes required to use remote EJB and JMS, including all dependencies. It is intended for use by those not using maven, maven users should just import the EJB and JMS BOM's instead (shaded JAR's cause lots of problems with maven, as it is very easy to inadvertently end up with different versions on classes on the class path).

There is a newer version: 34.0.0.Final
Show newest version
package org.jgroups.protocols;

import org.jgroups.*;
import org.jgroups.annotations.MBean;
import org.jgroups.annotations.ManagedAttribute;
import org.jgroups.annotations.ManagedOperation;
import org.jgroups.annotations.Property;
import org.jgroups.conf.AttributeType;
import org.jgroups.stack.Protocol;
import org.jgroups.util.*;

import java.util.*;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.ConcurrentMap;
import java.util.concurrent.Future;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.atomic.AtomicBoolean;
import java.util.concurrent.atomic.AtomicInteger;
import java.util.concurrent.atomic.AtomicLong;
import java.util.concurrent.atomic.LongAdder;
import java.util.concurrent.locks.ReentrantLock;
import java.util.function.BiConsumer;
import java.util.function.Predicate;
import java.util.function.Supplier;
import java.util.function.ToIntFunction;
import java.util.stream.Stream;

import static org.jgroups.Message.TransientFlag.DONT_LOOPBACK;
import static org.jgroups.Message.TransientFlag.OOB_DELIVERED;


/**
 * Reliable unicast protocol using a combination of positive and negative acks. See docs/design/UNICAST3.txt for details.
 * @author Bela Ban
 * @since  3.3
 */
@MBean(description="Reliable unicast layer")
public class UNICAST3 extends Protocol implements AgeOutCache.Handler
{ protected static final long DEFAULT_FIRST_SEQNO=Global.DEFAULT_FIRST_UNICAST_SEQNO; /* ------------------------------------------ Properties ------------------------------------------ */ @Property(description="Time (in milliseconds) after which an idle incoming or outgoing connection is closed. The " + "connection will get re-established when used again. 0 disables connection reaping. Note that this creates " + "lingering connection entries, which increases memory over time.",type=AttributeType.TIME) protected long conn_expiry_timeout=(long) 60000 * 2; @Property(description="Time (in ms) until a connection marked to be closed will get removed. 0 disables this", type=AttributeType.TIME) protected long conn_close_timeout=240_000; // 4 mins == TIME_WAIT timeout (= 2 * MSL) @Property(description="Number of rows of the matrix in the retransmission table (only for experts)",writable=false) protected int xmit_table_num_rows=100; @Property(description="Number of elements of a row of the matrix in the retransmission table; " + "gets rounded to the next power of 2 (only for experts). The capacity of the matrix is xmit_table_num_rows * xmit_table_msgs_per_row",writable=false) protected int xmit_table_msgs_per_row=1024; @Property(description="Resize factor of the matrix in the retransmission table (only for experts)",writable=false) protected double xmit_table_resize_factor=1.2; @Property(description="Number of milliseconds after which the matrix in the retransmission table " + "is compacted (only for experts)",writable=false,type=AttributeType.TIME) protected long xmit_table_max_compaction_time= (long) 10 * 60 * 1000; // @Property(description="Max time (in ms) after which a connection to a non-member is closed") protected long max_retransmit_time=60 * 1000L; @Property(description="Interval (in milliseconds) at which messages in the send windows are resent",type=AttributeType.TIME) protected long xmit_interval=500; @Property(description="If true, trashes warnings about retransmission messages not found in the xmit_table (used for testing)") protected boolean log_not_found_msgs=true; @Property(description="Send an ack immediately when a batch of ack_threshold (or more) messages is received. " + "Otherwise send delayed acks. If 1, ack single messages (similar to UNICAST)") protected int ack_threshold=100; @Property(description="Min time (in ms) to elapse for successive SEND_FIRST_SEQNO messages to be sent to the same sender", type=AttributeType.TIME) protected long sync_min_interval=2000; @Property(description="Max number of messages to ask for in a retransmit request. 0 disables this and uses " + "the max bundle size in the transport") protected int max_xmit_req_size; @Property(description="The max size of a message batch when delivering messages. 0 is unbounded") protected int max_batch_size; @Property(description="If true, a unicast message to self is looped back up on the same thread. Noter that this may " + "cause problems (e.g. deadlocks) in some applications, so make sure that your code can handle this. " + "Issue: https://issues.redhat.com/browse/JGRP-2547") protected boolean loopback; /* --------------------------------------------- JMX ---------------------------------------------- */ protected long num_msgs_sent=0, num_msgs_received=0; protected long num_acks_sent=0, num_acks_received=0, num_xmits=0; @ManagedAttribute(description="Number of retransmit requests received",type=AttributeType.SCALAR) protected final LongAdder xmit_reqs_received=new LongAdder(); @ManagedAttribute(description="Number of retransmit requests sent",type=AttributeType.SCALAR) protected final LongAdder xmit_reqs_sent=new LongAdder(); @ManagedAttribute(description="Number of retransmit responses sent",type=AttributeType.SCALAR) protected final LongAdder xmit_rsps_sent=new LongAdder(); @ManagedAttribute(description="Number of unicast messages to self looped back up",type=AttributeType.SCALAR) public long getNumLoopbacks() {return loopbed_back_msgs.sum();} @ManagedAttribute(description="Average batch size of messages delivered to the application") protected final AverageMinMax avg_delivery_batch_size=new AverageMinMax(); @ManagedAttribute(description="True if sending a message can block at the transport level") protected boolean sends_can_block=true; @ManagedAttribute(description="tracing is enabled or disabled for the given log",writable=true) protected boolean is_trace=log.isTraceEnabled(); /* --------------------------------------------- Fields ------------------------------------------------ */ protected final ConcurrentMap send_table=Util.createConcurrentMap(); protected final ConcurrentMap recv_table=Util.createConcurrentMap(); protected final ReentrantLock recv_table_lock=new ReentrantLock(); /** Used by the retransmit task to keep the last retransmitted seqno per sender (https://issues.redhat.com/browse/JGRP-1539) */ protected final Map xmit_task_map=new HashMap<>(); /** RetransmitTask running every xmit_interval ms */ protected Future xmit_task; protected volatile List
members=new ArrayList<>(11); protected TimeScheduler timer; // used for retransmissions protected volatile boolean running=false; protected short last_conn_id; protected AgeOutCache
cache; protected TimeService time_service; // for aging out of receiver and send entries protected final AtomicInteger timestamper=new AtomicInteger(0); // timestamping of ACKs / SEND_FIRST-SEQNOs /** Keep track of when a SEND_FIRST_SEQNO message was sent to a given sender */ protected ExpiryCache
last_sync_sent; protected final LongAdder loopbed_back_msgs=new LongAdder(); protected final MessageCache msg_cache=new MessageCache(); protected static final Message DUMMY_OOB_MSG=new EmptyMessage().setFlag(Message.Flag.OOB); protected final Predicate drop_oob_and_dont_loopback_msgs_filter=msg -> msg != null && msg != DUMMY_OOB_MSG && (!msg.isFlagSet(Message.Flag.OOB) || msg.setFlagIfAbsent(Message.TransientFlag.OOB_DELIVERED)) && !(msg.isFlagSet(DONT_LOOPBACK) && Objects.equals(local_addr, msg.getSrc())); protected static final Predicate dont_loopback_filter=m -> m != null && (m.isFlagSet(DONT_LOOPBACK) || m == DUMMY_OOB_MSG || m.isFlagSet(OOB_DELIVERED)); protected static final BiConsumer BATCH_ACCUMULATOR=MessageBatch::add; protected static final Table.Visitor DECR=(seqno, msg, row, col) -> { if(msg instanceof Refcountable) ((Refcountable)msg).decr(); return true; }; @ManagedAttribute(description="Returns the number of outgoing (send) connections") public int getNumSendConnections() { return send_table.size(); } @ManagedAttribute(description="Returns the number of incoming (receive) connections") public int getNumReceiveConnections() { return recv_table.size(); } @ManagedAttribute(description="Returns the total number of outgoing (send) and incoming (receive) connections") public int getNumConnections() { return getNumReceiveConnections() + getNumSendConnections(); } @ManagedAttribute(description="Next seqno issued by the timestamper") public int getTimestamper() {return timestamper.get();} public int getAckThreshold() {return ack_threshold;} public UNICAST3 setAckThreshold(int a) {ack_threshold=a; return this;} @Property(name="level", description="Sets the level") public T setLevel(String level) { T retval= super.setLevel(level); is_trace=log.isTraceEnabled(); return retval; } public long getXmitInterval() {return xmit_interval;} public UNICAST3 setXmitInterval(long i) {xmit_interval=i; return this;} public int getXmitTableNumRows() {return xmit_table_num_rows;} public UNICAST3 setXmitTableNumRows(int n) {this.xmit_table_num_rows=n; return this;} public int getXmitTableMsgsPerRow() {return xmit_table_msgs_per_row;} public UNICAST3 setXmitTableMsgsPerRow(int n) {this.xmit_table_msgs_per_row=n; return this;} public long getConnExpiryTimeout() {return conn_expiry_timeout;} public UNICAST3 setConnExpiryTimeout(long c) {this.conn_expiry_timeout=c; return this;} public long getConnCloseTimeout() {return conn_close_timeout;} public UNICAST3 setConnCloseTimeout(long c) {this.conn_close_timeout=c; return this;} public double getXmitTableResizeFactor() {return xmit_table_resize_factor;} public UNICAST3 setXmitTableResizeFactor(double x) {this.xmit_table_resize_factor=x; return this;} public long getXmitTableMaxCompactionTime() {return xmit_table_max_compaction_time;} public UNICAST3 setXmitTableMaxCompactionTime(long x) {this.xmit_table_max_compaction_time=x; return this;} public boolean logNotFoundMsgs() {return log_not_found_msgs;} public UNICAST3 logNotFoundMsgs(boolean l) {this.log_not_found_msgs=l; return this;} public long getSyncMinInterval() {return sync_min_interval;} public UNICAST3 setSyncMinInterval(long s) {this.sync_min_interval=s; return this;} public int getMaxXmitReqSize() {return max_xmit_req_size;} public UNICAST3 setMaxXmitReqSize(int m) {this.max_xmit_req_size=m; return this;} public boolean sendsCanBlock() {return sends_can_block;} public UNICAST3 sendsCanBlock(boolean s) {this.sends_can_block=s; return this;} @ManagedOperation public String printConnections() { StringBuilder sb=new StringBuilder(); if(!send_table.isEmpty()) { sb.append("\nsend connections:\n"); for(Map.Entry entry: send_table.entrySet()) { sb.append(entry.getKey()).append(": ").append(entry.getValue()).append("\n"); } } if(!recv_table.isEmpty()) { sb.append("\nreceive connections:\n"); for(Map.Entry entry: recv_table.entrySet()) { sb.append(entry.getKey()).append(": ").append(entry.getValue()).append("\n"); } } return sb.toString(); } @ManagedAttribute(type=AttributeType.SCALAR) public long getNumMessagesSent() {return num_msgs_sent;} @ManagedAttribute(type=AttributeType.SCALAR) public long getNumMessagesReceived() {return num_msgs_received;} @ManagedAttribute(type=AttributeType.SCALAR) public long getNumAcksSent() {return num_acks_sent;} @ManagedAttribute(type=AttributeType.SCALAR) public long getNumAcksReceived() {return num_acks_received;} @ManagedAttribute(type=AttributeType.SCALAR) public long getNumXmits() {return num_xmits;} public long getMaxRetransmitTime() {return max_retransmit_time;} @Property(description="Max number of milliseconds we try to retransmit a message to any given member. After that, " + "the connection is removed. Any new connection to that member will start with seqno #1 again. 0 disables this", type=AttributeType.TIME) public UNICAST3 setMaxRetransmitTime(long max_retransmit_time) { this.max_retransmit_time=max_retransmit_time; if(cache != null && max_retransmit_time > 0) cache.setTimeout(max_retransmit_time); return this; } @ManagedAttribute(description="Is the retransmit task running") public boolean isXmitTaskRunning() {return xmit_task != null && !xmit_task.isDone();} @ManagedAttribute public int getAgeOutCacheSize() { return cache != null? cache.size() : 0; } @ManagedOperation public String printAgeOutCache() { return cache != null? cache.toString() : "n/a"; } public AgeOutCache
getAgeOutCache() { return cache; } /** Used for testing only */ public boolean hasSendConnectionTo(Address dest) { Entry entry=send_table.get(dest); return entry != null && entry.state() == State.OPEN; } /** The number of messages in all Entry.sent_msgs tables (haven't received an ACK yet) */ @ManagedAttribute(type=AttributeType.SCALAR) public int getNumUnackedMessages() { return accumulate(Table::size, send_table.values()); } @ManagedAttribute(description="Total number of undelivered messages in all receive windows",type=AttributeType.SCALAR) public int getXmitTableUndeliveredMessages() { return accumulate(Table::size, recv_table.values()); } @ManagedAttribute(description="Total number of missing messages in all receive windows",type=AttributeType.SCALAR) public int getXmitTableMissingMessages() { return accumulate(Table::getNumMissing, recv_table.values()); } @ManagedAttribute(description="Total number of deliverable messages in all receive windows",type=AttributeType.SCALAR) public int getXmitTableDeliverableMessages() { return accumulate(Table::getNumDeliverable, recv_table.values()); } @ManagedAttribute(description="Number of compactions in all (receive and send) windows") public int getXmitTableNumCompactions() { return accumulate(Table::getNumCompactions, recv_table.values(), send_table.values()); } @ManagedAttribute(description="Number of moves in all (receive and send) windows") public int getXmitTableNumMoves() { return accumulate(Table::getNumMoves, recv_table.values(), send_table.values()); } @ManagedAttribute(description="Number of resizes in all (receive and send) windows") public int getXmitTableNumResizes() { return accumulate(Table::getNumResizes, recv_table.values(), send_table.values()); } @ManagedAttribute(description="Number of purges in all (receive and send) windows") public int getXmitTableNumPurges() { return accumulate(Table::getNumPurges, recv_table.values(), send_table.values()); } @ManagedOperation(description="Prints the contents of the receive windows for all members") public String printReceiveWindowMessages() { StringBuilder ret=new StringBuilder(local_addr + ":\n"); for(Map.Entry entry: recv_table.entrySet()) { Address addr=entry.getKey(); Table buf=entry.getValue().msgs; ret.append(addr).append(": ").append(buf.toString()).append('\n'); } return ret.toString(); } @ManagedOperation(description="Prints the contents of the send windows for all members") public String printSendWindowMessages() { StringBuilder ret=new StringBuilder(local_addr + ":\n"); for(Map.Entry entry: send_table.entrySet()) { Address addr=entry.getKey(); Table buf=entry.getValue().msgs; ret.append(addr).append(": ").append(buf.toString()).append('\n'); } return ret.toString(); } public void resetStats() { num_msgs_sent=num_msgs_received=num_acks_sent=num_acks_received=num_xmits=0; avg_delivery_batch_size.clear(); Stream.of(xmit_reqs_received, xmit_reqs_sent, xmit_rsps_sent).forEach(LongAdder::reset); loopbed_back_msgs.reset(); } public void init() throws Exception { super.init(); TP transport=getTransport(); sends_can_block=transport instanceof TCP; // UDP and TCP_NIO2 won't block time_service=transport.getTimeService(); if(time_service == null) throw new IllegalStateException("time service from transport is null"); last_sync_sent=new ExpiryCache<>(sync_min_interval); // max bundle size (minus overhead) divided by times bits per long // Example: for 8000 missing messages, SeqnoList has a serialized size of 1012 bytes, for 64000 messages, the // serialized size is 8012 bytes. Therefore, for a serialized size of 64000 bytes, we can retransmit a max of // 8 * 64000 = 512'000 seqnos // see SeqnoListTest.testSerialization3() int estimated_max_msgs_in_xmit_req=(transport.getBundler().getMaxSize() -50) * Global.LONG_SIZE; int old_max_xmit_size=max_xmit_req_size; if(max_xmit_req_size <= 0) max_xmit_req_size=estimated_max_msgs_in_xmit_req; else max_xmit_req_size=Math.min(max_xmit_req_size, estimated_max_msgs_in_xmit_req); if(old_max_xmit_size != max_xmit_req_size) log.trace("%s: set max_xmit_req_size from %d to %d", local_addr, old_max_xmit_size, max_xmit_req_size); } public void start() throws Exception { msg_cache.clear(); timer=getTransport().getTimer(); if(timer == null) throw new Exception("timer is null"); if(max_retransmit_time > 0) cache=new AgeOutCache<>(timer, max_retransmit_time, this); running=true; startRetransmitTask(); } public void stop() { sendPendingAcks(); running=false; stopRetransmitTask(); xmit_task_map.clear(); removeAllConnections(); msg_cache.clear(); } public Object up(Message msg) { if(msg.getDest() == null || msg.isFlagSet(Message.Flag.NO_RELIABILITY)) // only handle unicast messages return up_prot.up(msg); // pass up UnicastHeader3 hdr=msg.getHeader(this.id); if(hdr == null) return up_prot.up(msg); Address sender=msg.getSrc(); switch(hdr.type) { case UnicastHeader3.DATA: // received regular message if(is_trace) log.trace("%s <-- %s: DATA(#%d, conn_id=%d%s)", local_addr, sender, hdr.seqno, hdr.conn_id, hdr.first? ", first" : ""); if(Objects.equals(local_addr, sender)) handleDataReceivedFromSelf(sender, hdr.seqno, msg); else handleDataReceived(sender, hdr.seqno, hdr.conn_id, hdr.first, msg); break; // we pass the deliverable message up in handleDataReceived() default: handleUpEvent(sender, msg, hdr); break; } return null; } protected void handleUpEvent(Address sender, Message msg, UnicastHeader3 hdr) { try { switch(hdr.type) { case UnicastHeader3.DATA: // received regular message throw new IllegalStateException("header of type DATA is not supposed to be handled by this method"); case UnicastHeader3.ACK: // received ACK for previously sent message handleAckReceived(sender, hdr.seqno, hdr.conn_id, hdr.timestamp()); break; case UnicastHeader3.SEND_FIRST_SEQNO: handleResendingOfFirstMessage(sender, hdr.timestamp()); break; case UnicastHeader3.XMIT_REQ: // received ACK for previously sent message handleXmitRequest(sender, msg.getObject()); break; case UnicastHeader3.CLOSE: log.trace("%s <-- %s: CLOSE(conn-id=%s)", local_addr, sender, hdr.conn_id); ReceiverEntry entry=recv_table.get(sender); if(entry != null && entry.connId() == hdr.conn_id) { recv_table.remove(sender, entry); log.trace("%s: removed receive connection for %s", local_addr, sender); } break; default: log.error(Util.getMessage("TypeNotKnown"), local_addr, hdr.type); break; } } catch(Throwable t) { // we cannot let an exception terminate the processing of this batch log.error(Util.getMessage("FailedHandlingEvent"), local_addr, t); } } public void up(MessageBatch batch) { if(batch.dest() == null) { // not a unicast batch up_prot.up(batch); return; } final Address sender=batch.sender(); if(local_addr == null || local_addr.equals(sender)) { Entry entry=local_addr != null? send_table.get(local_addr) : null; if(entry != null) handleBatchFromSelf(batch, entry); return; } int size=batch.size(); Map>> msgs=new LinkedHashMap<>(); ReceiverEntry entry=recv_table.get(sender); for(Iterator it=batch.iterator(); it.hasNext();) { Message msg=it.next(); UnicastHeader3 hdr; if(msg == null || msg.isFlagSet(Message.Flag.NO_RELIABILITY) || (hdr=msg.getHeader(id)) == null) continue; it.remove(); // remove the message from the batch, so it won't be passed up the stack if(hdr.type != UnicastHeader3.DATA) { handleUpEvent(msg.getSrc(), msg, hdr); continue; } List> list=msgs.computeIfAbsent(hdr.conn_id, k -> new ArrayList<>(size)); list.add(new LongTuple<>(hdr.seqno(), msg)); if(hdr.first) entry=getReceiverEntry(sender, hdr.seqno(), hdr.first, hdr.connId()); else if(entry == null) { msg_cache.cache(sender, msg); log.trace("%s: cached %s#%d", local_addr, sender, hdr.seqno()); } } if(!msgs.isEmpty()) { if(entry == null) sendRequestForFirstSeqno(sender); else { if(!msg_cache.isEmpty()) { // quick and dirty check List queued_msgs=msg_cache.drain(sender); if(queued_msgs != null) addQueuedMessages(sender, entry, queued_msgs); } if(msgs.keySet().retainAll(Collections.singletonList(entry.connId()))) // remove all conn-ids that don't match sendRequestForFirstSeqno(sender); List> list=msgs.get(entry.connId()); if(list != null && !list.isEmpty()) handleBatchReceived(entry, sender, list, batch.mode() == MessageBatch.Mode.OOB); } } if(!batch.isEmpty()) up_prot.up(batch); } protected void handleBatchFromSelf(MessageBatch batch, Entry entry) { List> list=new ArrayList<>(batch.size()); for(Iterator it=batch.iterator(); it.hasNext();) { Message msg=it.next(); UnicastHeader3 hdr; if(msg == null || msg.isFlagSet(Message.Flag.NO_RELIABILITY) || (hdr=msg.getHeader(id)) == null) continue; it.remove(); // remove the message from the batch, so it won't be passed up the stack if(hdr.type != UnicastHeader3.DATA) { handleUpEvent(msg.getSrc(), msg, hdr); continue; } if(entry.conn_id != hdr.conn_id) { it.remove(); continue; } list.add(new LongTuple<>(hdr.seqno(), msg)); } if(!list.isEmpty()) { if(is_trace) log.trace("%s <-- %s: DATA(%s)", local_addr, batch.sender(), printMessageList(list)); int len=list.size(); Table win=entry.msgs; update(entry, len); // OOB msg is passed up. When removed, we discard it. Affects ordering: https://issues.redhat.com/browse/JGRP-379 if(batch.mode() == MessageBatch.Mode.OOB) { MessageBatch oob_batch=new MessageBatch(local_addr, batch.sender(), batch.clusterName(), batch.multicast(), MessageBatch.Mode.OOB, len); for(LongTuple tuple: list) { long seq=tuple.getVal1(); Message msg=win.get(seq); // we *have* to get the message, because loopback means we didn't add it to win ! if(msg != null && msg.isFlagSet(Message.Flag.OOB) && msg.setFlagIfAbsent(Message.TransientFlag.OOB_DELIVERED)) oob_batch.add(msg); } deliverBatch(oob_batch); } removeAndDeliver(win, batch.sender()); } if(!batch.isEmpty()) up_prot.up(batch); } public Object down(Event evt) { switch (evt.getType()) { case Event.VIEW_CHANGE: // remove connections to peers that are not members anymore ! View view=evt.getArg(); List
new_members=view.getMembers(); Set
non_members=new HashSet<>(send_table.keySet()); non_members.addAll(recv_table.keySet()); members=new_members; new_members.forEach(non_members::remove); if(cache != null) cache.removeAll(new_members); if(!non_members.isEmpty()) { log.trace("%s: closing connections of non members %s", local_addr, non_members); non_members.forEach(this::closeConnection); } if(!new_members.isEmpty()) { for(Address mbr: new_members) { Entry e=send_table.get(mbr); if(e != null && e.state() == State.CLOSING) e.state(State.OPEN); e=recv_table.get(mbr); if(e != null && e.state() == State.CLOSING) e.state(State.OPEN); } } xmit_task_map.keySet().retainAll(new_members); last_sync_sent.removeExpiredElements(); break; } return down_prot.down(evt); // Pass on to the layer below us } public Object down(Message msg) { Address dst=msg.getDest(); /* only handle unicast messages */ if (dst == null || msg.isFlagSet(Message.Flag.NO_RELIABILITY)) return down_prot.down(msg); if(!running) { log.trace("%s: discarded message as start() has not yet been called, message: %s", local_addr, msg); return null; } if(msg.getSrc() == null) msg.setSrc(local_addr); // this needs to be done so we can check whether the message sender is the local_addr if(loopback && Objects.equals(local_addr, dst)) {// https://issues.redhat.com/browse/JGRP-2547 if(msg.isFlagSet(DONT_LOOPBACK)) return null; loopbed_back_msgs.increment(); return up_prot.up(msg); } SenderEntry entry=getSenderEntry(dst); boolean dont_loopback_set=msg.isFlagSet(DONT_LOOPBACK) && dst.equals(local_addr); short send_conn_id=entry.connId(); long seqno=entry.sent_msgs_seqno.getAndIncrement(); long sleep=10; if(msg instanceof Refcountable && !dont_loopback_set) ((Refcountable)msg).incr(); do { try { msg.putHeader(this.id,UnicastHeader3.createDataHeader(seqno,send_conn_id,seqno == DEFAULT_FIRST_SEQNO)); // add *including* UnicastHeader, adds to retransmitter entry.msgs.add(seqno, msg, dont_loopback_set? dont_loopback_filter : null); if(conn_expiry_timeout > 0) entry.update(); if(dont_loopback_set) entry.msgs.purge(entry.msgs.getHighestDeliverable()); break; } catch(Throwable t) { if(running) { Util.sleep(sleep); sleep=Math.min(5000, sleep*2); } } } while(running); if(is_trace) { StringBuilder sb=new StringBuilder(); sb.append(local_addr).append(" --> ").append(dst).append(": DATA(").append("#").append(seqno). append(", conn_id=").append(send_conn_id); if(seqno == DEFAULT_FIRST_SEQNO) sb.append(", first"); sb.append(')'); log.trace(sb); } num_msgs_sent++; return down_prot.down(msg); } /** * Removes and resets from connection table (which is already locked). Returns true if member was found, * otherwise false. This method is public only so it can be invoked by unit testing, but should not be used ! */ public void closeConnection(Address mbr) { closeSendConnection(mbr); closeReceiveConnection(mbr); } public void closeSendConnection(Address mbr) { SenderEntry entry=send_table.get(mbr); if(entry != null) entry.state(State.CLOSING); } public void closeReceiveConnection(Address mbr) { ReceiverEntry entry=recv_table.get(mbr); if(entry != null) entry.state(State.CLOSING); } protected void removeSendConnection(Address mbr) { SenderEntry entry=send_table.remove(mbr); if(entry != null) { entry.state(State.CLOSED); if(members.contains(mbr)) sendClose(mbr, entry.connId()); } } protected void removeReceiveConnection(Address mbr) { sendPendingAcks(); ReceiverEntry entry=recv_table.remove(mbr); if(entry != null) entry.state(State.CLOSED); } /** * This method is public only so it can be invoked by unit testing, but should not otherwise be used ! */ @ManagedOperation(description="Trashes all connections to other nodes. This is only used for testing") public void removeAllConnections() { send_table.clear(); recv_table.clear(); } /** Sends a retransmit request to the given sender */ protected void retransmit(SeqnoList missing, Address sender) { Message xmit_msg=new ObjectMessage(sender, missing).setFlag(Message.Flag.OOB) .putHeader(id, UnicastHeader3.createXmitReqHeader()); if(is_trace) log.trace("%s --> %s: XMIT_REQ(%s)", local_addr, sender, missing); down_prot.down(xmit_msg); xmit_reqs_sent.add(missing.size()); } /** Called by the sender to resend messages for which no ACK has been received yet */ protected void retransmit(Message msg) { if(is_trace) { UnicastHeader3 hdr=msg.getHeader(id); long seqno=hdr != null? hdr.seqno : -1; log.trace("%s --> %s: resending(#%d)", local_addr, msg.getDest(), seqno); } down_prot.down(msg); num_xmits++; } /** * Called by AgeOutCache, to removed expired connections * @param key */ public void expired(Address key) { if(key != null) { log.debug("%s: removing expired connection to %s", local_addr, key); closeConnection(key); } } /** * Check whether the hashtable contains an entry e for {@code sender} (create if not). If * e.received_msgs is null and {@code first} is true: create a new AckReceiverWindow(seqno) and * add message. Set e.received_msgs to the new window. Else just add the message. */ protected void handleDataReceived(final Address sender, long seqno, short conn_id, boolean first, final Message msg) { ReceiverEntry entry=getReceiverEntry(sender, seqno, first, conn_id); if(entry == null) { msg_cache.cache(sender, msg); log.trace("%s: cached %s#%d", local_addr, sender, seqno); return; } if(!msg_cache.isEmpty()) { // quick and dirty check List queued_msgs=msg_cache.drain(sender); if(queued_msgs != null) addQueuedMessages(sender, entry, queued_msgs); } addMessage(entry, sender, seqno, msg); removeAndDeliver(entry.msgs, sender); } protected void addMessage(ReceiverEntry entry, Address sender, long seqno, Message msg) { final Table win=entry.msgs; update(entry, 1); boolean oob=msg.isFlagSet(Message.Flag.OOB), added=win.add(seqno, oob? DUMMY_OOB_MSG : msg); // adding the same dummy OOB msg saves space (we won't remove it) if(ack_threshold <= 1) sendAck(sender, win.getHighestDeliverable(), entry.connId()); else entry.sendAck(); // will be sent delayed (on the next xmit_interval) // An OOB message is passed up immediately. Later, when remove() is called, we discard it. This affects ordering ! // https://issues.redhat.com/browse/JGRP-377 if(oob && added) deliverMessage(msg, sender, seqno); } protected void addQueuedMessages(final Address sender, final ReceiverEntry entry, List queued_msgs) { for(Message msg: queued_msgs) { UnicastHeader3 hdr=msg.getHeader(this.id); if(hdr.conn_id != entry.conn_id) { log.warn("%s: dropped queued message %s#%d as its conn_id (%d) did not match (entry.conn_id=%d)", local_addr, sender, hdr.seqno, hdr.conn_id, entry.conn_id); continue; } addMessage(entry, sender, hdr.seqno(), msg); } } /** Called when the sender of a message is the local member. In this case, we don't need to add the message * to the table as the sender already did that */ protected void handleDataReceivedFromSelf(final Address sender, long seqno, Message msg) { Entry entry=send_table.get(sender); if(entry == null || entry.state() == State.CLOSED) { log.warn("%s: entry not found for %s; dropping message", local_addr, sender); return; } update(entry, 1); final Table win=entry.msgs; // An OOB message is passed up immediately. Later, when remove() is called, we discard it. // This affects ordering ! JIRA: https://issues.redhat.com/browse/JGRP-377 if(msg.isFlagSet(Message.Flag.OOB)) { msg=win.get(seqno); // we *have* to get a message, because loopback means we didn't add it to win ! if(msg != null && msg.isFlagSet(Message.Flag.OOB) && msg.setFlagIfAbsent(Message.TransientFlag.OOB_DELIVERED)) deliverMessage(msg, sender, seqno); } removeAndDeliver(win, sender); // there might be more messages to deliver } protected void handleBatchReceived(final ReceiverEntry entry, Address sender, List> msgs, boolean oob) { if(is_trace) log.trace("%s <-- %s: DATA(%s)", local_addr, sender, printMessageList(msgs)); int batch_size=msgs.size(); Table win=entry.msgs; // adds all messages to the table, removing messages from 'msgs' which could not be added (already present) boolean added=win.add(msgs, oob, oob? DUMMY_OOB_MSG : null); update(entry, batch_size); if(batch_size >= ack_threshold) sendAck(sender, win.getHighestDeliverable(), entry.connId()); else entry.sendAck(); // OOB msg is passed up. When removed, we discard it. Affects ordering: https://issues.redhat.com/browse/JGRP-379 if(added && oob) { MessageBatch oob_batch=new MessageBatch(local_addr, sender, null, false, MessageBatch.Mode.OOB, msgs.size()); for(LongTuple tuple: msgs) oob_batch.add(tuple.getVal2()); deliverBatch(oob_batch); } removeAndDeliver(win, sender); } /** * Try to remove as many messages as possible from the table as pass them up. * Prevents concurrent passing up of messages by different threads (https://issues.redhat.com/browse/JGRP-198); * lots of threads can come up to this point concurrently, but only 1 is allowed to pass at a time. * We *can* deliver messages from *different* senders concurrently, e.g. reception of P1, Q1, P2, Q2 can result in * delivery of P1, Q1, Q2, P2: FIFO (implemented by UNICAST) says messages need to be delivered in the * order in which they were sent */ protected void removeAndDeliver(Table win, Address sender) { AtomicInteger adders=win.getAdders(); if(adders.getAndIncrement() != 0) return; final MessageBatch batch=new MessageBatch(win.getNumDeliverable()) .dest(local_addr).sender(sender).multicast(false); Supplier batch_creator=() -> batch; do { try { batch.reset(); // sets index to 0: important as batch delivery may not remove messages from batch! win.removeMany(true, max_batch_size, drop_oob_and_dont_loopback_msgs_filter, batch_creator, BATCH_ACCUMULATOR); } catch(Throwable t) { log.error("%s: failed removing messages from table for %s: %s", local_addr, sender, t); } if(!batch.isEmpty()) { // batch is guaranteed to NOT contain any OOB messages as the drop_oob_msgs_filter above removed them deliverBatch(batch); // catches Throwable } } while(adders.decrementAndGet() != 0); } protected String printMessageList(List> list) { StringBuilder sb=new StringBuilder(); int size=list.size(); Message first=size > 0? list.get(0).getVal2() : null, second=size > 1? list.get(size-1).getVal2() : first; UnicastHeader3 hdr; if(first != null) { hdr=first.getHeader(id); if(hdr != null) sb.append("#" + hdr.seqno); } if(second != null) { hdr=second.getHeader(id); if(hdr != null) sb.append(" - #" + hdr.seqno); } return sb.toString(); } protected ReceiverEntry getReceiverEntry(Address sender, long seqno, boolean first, short conn_id) { ReceiverEntry entry=recv_table.get(sender); if(entry != null && entry.connId() == conn_id) return entry; recv_table_lock.lock(); try { entry=recv_table.get(sender); if(first) { if(entry == null) { entry=createReceiverEntry(sender,seqno,conn_id); } else { // entry != null && win != null if(conn_id != entry.connId()) { log.trace("%s: conn_id=%d != %d; resetting receiver window", local_addr, conn_id, entry.connId()); recv_table.remove(sender); entry=createReceiverEntry(sender,seqno,conn_id); } } } else { // entry == null && win == null OR entry != null && win == null OR entry != null && win != null if(entry == null || entry.connId() != conn_id) { recv_table_lock.unlock(); sendRequestForFirstSeqno(sender); // drops the message and returns (see below) return null; } } return entry; } finally { if(recv_table_lock.isHeldByCurrentThread()) recv_table_lock.unlock(); } } protected SenderEntry getSenderEntry(Address dst) { SenderEntry entry=send_table.get(dst); if(entry == null || entry.state() == State.CLOSED) { if(entry != null) send_table.remove(dst, entry); entry=send_table.computeIfAbsent(dst, k -> new SenderEntry(getNewConnectionId())); log.trace("%s: created sender window for %s (conn-id=%s)", local_addr, dst, entry.connId()); if(cache != null && !members.contains(dst)) cache.add(dst); } if(entry.state() == State.CLOSING) entry.state(State.OPEN); return entry; } protected ReceiverEntry createReceiverEntry(Address sender, long seqno, short conn_id) { ReceiverEntry entry=recv_table.computeIfAbsent(sender, k -> new ReceiverEntry(createTable(seqno), conn_id)); log.trace("%s: created receiver window for %s at seqno=#%d for conn-id=%d", local_addr, sender, seqno, conn_id); return entry; } protected Table createTable(long seqno) { return new Table<>(xmit_table_num_rows, xmit_table_msgs_per_row, seqno-1, xmit_table_resize_factor, xmit_table_max_compaction_time); } /** Add the ACK to hashtable.sender.sent_msgs */ protected void handleAckReceived(Address sender, long seqno, short conn_id, int timestamp) { if(is_trace) log.trace("%s <-- %s: ACK(#%d, conn-id=%d, ts=%d)", local_addr, sender, seqno, conn_id, timestamp); SenderEntry entry=send_table.get(sender); if(entry != null && entry.connId() != conn_id) { log.trace("%s: my conn_id (%d) != received conn_id (%d); discarding ACK", local_addr, entry.connId(), conn_id); return; } Table win=entry != null? entry.msgs : null; if(win != null && entry.updateLastTimestamp(timestamp)) { win.forEach(win.getLow(), seqno, DECR); win.purge(seqno, true); // removes all messages <= seqno (forced purge) num_acks_received++; } } /** * We need to resend the first message with our conn_id * @param sender */ protected void handleResendingOfFirstMessage(Address sender, int timestamp) { log.trace("%s <-- %s: SEND_FIRST_SEQNO", local_addr, sender); SenderEntry entry=send_table.get(sender); Table win=entry != null? entry.msgs : null; if(win == null) { log.warn(Util.getMessage("SenderNotFound"), local_addr, sender); return; } if(!entry.updateLastTimestamp(timestamp)) return; Message rsp=win.get(win.getLow() +1); if(rsp != null) { // We need to copy the UnicastHeader and put it back into the message because Message.copy() doesn't copy // the headers and therefore we'd modify the original message in the sender retransmission window // (https://issues.redhat.com/browse/JGRP-965) Message copy=rsp.copy(true, true); UnicastHeader3 hdr=copy.getHeader(this.id); UnicastHeader3 newhdr=hdr.copy(); newhdr.first=true; copy.putHeader(this.id, newhdr); down_prot.down(copy); } } protected void handleXmitRequest(Address sender, SeqnoList missing) { if(is_trace) log.trace("%s <-- %s: XMIT(#%s)", local_addr, sender, missing); SenderEntry entry=send_table.get(sender); xmit_reqs_received.add(missing.size()); Table win=entry != null? entry.msgs : null; if(win != null) { for(Long seqno: missing) { Message msg=win.get(seqno); if(msg == null) { if(log.isWarnEnabled() && log_not_found_msgs && !local_addr.equals(sender) && seqno > win.getLow()) log.warn(Util.getMessage("MessageNotFound"), local_addr, sender, seqno); continue; } down_prot.down(msg); xmit_rsps_sent.increment(); } } } protected void deliverMessage(final Message msg, final Address sender, final long seqno) { if(is_trace) log.trace("%s: delivering %s#%s", local_addr, sender, seqno); try { up_prot.up(msg); } catch(Throwable t) { log.warn(Util.getMessage("FailedToDeliverMsg"), local_addr, msg.isFlagSet(Message.Flag.OOB) ? "OOB message" : "message", msg, t); } } protected void deliverBatch(MessageBatch batch) { try { if(batch.isEmpty()) return; if(is_trace) { Message first=batch.first(), last=batch.last(); StringBuilder sb=new StringBuilder(local_addr + ": delivering"); if(first != null && last != null) { UnicastHeader3 hdr1=first.getHeader(id), hdr2=last.getHeader(id); if(hdr1 != null && hdr2 != null) sb.append(" #").append(hdr1.seqno).append(" - #").append(hdr2.seqno); } sb.append(" (" + batch.size()).append(" messages)"); log.trace(sb); } if(stats) avg_delivery_batch_size.add(batch.size()); up_prot.up(batch); } catch(Throwable t) { log.warn(Util.getMessage("FailedToDeliverMsg"), local_addr, "batch", batch, t); } } protected long getTimestamp() { return time_service.timestamp(); } protected void startRetransmitTask() { if(xmit_task == null || xmit_task.isDone()) xmit_task=timer.scheduleWithFixedDelay(new RetransmitTask(), 0, xmit_interval, TimeUnit.MILLISECONDS, sends_can_block); } protected void stopRetransmitTask() { if(xmit_task != null) { xmit_task.cancel(true); xmit_task=null; } } protected void sendAck(Address dst, long seqno, short conn_id) { if(!running) // if we are disconnected, then don't send any acks which throw exceptions on shutdown return; Message ack=new EmptyMessage(dst). putHeader(this.id, UnicastHeader3.createAckHeader(seqno, conn_id, timestamper.incrementAndGet())); if(is_trace) log.trace("%s --> %s: ACK(#%d)", local_addr, dst, seqno); try { down_prot.down(ack); num_acks_sent++; } catch(Throwable t) { log.error(Util.getMessage("FailedSendingAck"), local_addr, seqno, dst, t); } } protected synchronized short getNewConnectionId() { short retval=last_conn_id; if(last_conn_id >= Short.MAX_VALUE || last_conn_id < 0) last_conn_id=0; else last_conn_id++; return retval; } protected void sendRequestForFirstSeqno(Address dest) { if(last_sync_sent.addIfAbsentOrExpired(dest)) { Message msg=new EmptyMessage(dest).setFlag(Message.Flag.OOB) .putHeader(this.id, UnicastHeader3.createSendFirstSeqnoHeader(timestamper.incrementAndGet())); log.trace("%s --> %s: SEND_FIRST_SEQNO", local_addr, dest); down_prot.down(msg); } } public void sendClose(Address dest, short conn_id) { Message msg=new EmptyMessage(dest).putHeader(id, UnicastHeader3.createCloseHeader(conn_id)); log.trace("%s --> %s: CLOSE(conn-id=%d)", local_addr, dest, conn_id); down_prot.down(msg); } @ManagedOperation(description="Closes connections that have been idle for more than conn_expiry_timeout ms") public void closeIdleConnections() { // close expired connections in send_table for(Map.Entry entry: send_table.entrySet()) { SenderEntry val=entry.getValue(); if(val.state() != State.OPEN) // only look at open connections continue; long age=val.age(); if(age >= conn_expiry_timeout) { log.debug("%s: closing expired connection for %s (%d ms old) in send_table", local_addr, entry.getKey(), age); closeSendConnection(entry.getKey()); } } // close expired connections in recv_table for(Map.Entry entry: recv_table.entrySet()) { ReceiverEntry val=entry.getValue(); if(val.state() != State.OPEN) // only look at open connections continue; long age=val.age(); if(age >= conn_expiry_timeout) { log.debug("%s: closing expired connection for %s (%d ms old) in recv_table", local_addr, entry.getKey(), age); closeReceiveConnection(entry.getKey()); } } } @ManagedOperation(description="Removes connections that have been closed for more than conn_close_timeout ms") public int removeExpiredConnections() { int num_removed=0; // remove expired connections from send_table for(Map.Entry entry: send_table.entrySet()) { SenderEntry val=entry.getValue(); if(val.state() == State.OPEN) // only look at closing or closed connections continue; long age=val.age(); if(age >= conn_close_timeout) { log.debug("%s: removing expired connection for %s (%d ms old) from send_table", local_addr, entry.getKey(), age); removeSendConnection(entry.getKey()); num_removed++; } } // remove expired connections from recv_table for(Map.Entry entry: recv_table.entrySet()) { ReceiverEntry val=entry.getValue(); if(val.state() == State.OPEN) // only look at closing or closed connections continue; long age=val.age(); if(age >= conn_close_timeout) { log.debug("%s: removing expired connection for %s (%d ms old) from recv_table", local_addr, entry.getKey(), age); removeReceiveConnection(entry.getKey()); num_removed++; } } return num_removed; } /** * Removes send- and/or receive-connections whose state is not OPEN (CLOSING or CLOSED). * @param remove_send_connections If true, send connections whose state is !OPEN are destroyed and removed * @param remove_receive_connections If true, receive connections with state !OPEN are destroyed and removed * @return The number of connections which were removed */ @ManagedOperation(description="Removes send- and/or receive-connections whose state is not OPEN (CLOSING or CLOSED)") public int removeConnections(boolean remove_send_connections, boolean remove_receive_connections) { int num_removed=0; if(remove_send_connections) { for(Map.Entry entry: send_table.entrySet()) { SenderEntry val=entry.getValue(); if(val.state() != State.OPEN) { // only look at closing or closed connections log.debug("%s: removing connection for %s (%d ms old, state=%s) from send_table", local_addr, entry.getKey(), val.age(), val.state()); removeSendConnection(entry.getKey()); num_removed++; } } } if(remove_receive_connections) { for(Map.Entry entry: recv_table.entrySet()) { ReceiverEntry val=entry.getValue(); if(val.state() != State.OPEN) { // only look at closing or closed connections log.debug("%s: removing expired connection for %s (%d ms old, state=%s) from recv_table", local_addr, entry.getKey(), val.age(), val.state()); removeReceiveConnection(entry.getKey()); num_removed++; } } } return num_removed; } @ManagedOperation(description="Triggers the retransmission task") public void triggerXmit() { SeqnoList missing; for(Map.Entry entry: recv_table.entrySet()) { Address target=entry.getKey(); // target to send retransmit requests to ReceiverEntry val=entry.getValue(); Table win=val != null? val.msgs : null; // receiver: send ack for received messages if needed if(win != null && val.needToSendAck()) // sendAck() resets send_ack to false sendAck(target, win.getHighestDeliverable(), val.connId()); // receiver: retransmit missing messages (getNumMissing() is fast) if(win != null && win.getNumMissing() > 0 && (missing=win.getMissing(max_xmit_req_size)) != null) { long highest=missing.getLast(); Long prev_seqno=xmit_task_map.get(target); if(prev_seqno == null) xmit_task_map.put(target, highest); // no retransmission else { missing.removeHigherThan(prev_seqno); // we only retransmit the 'previous batch' if(highest > prev_seqno) xmit_task_map.put(target, highest); if(!missing.isEmpty()) { // remove msgs that are <= highest-delivered (https://issues.redhat.com/browse/JGRP-2574) long highest_deliverable=win.getHighestDeliverable(), first=missing.getFirst(); if(first < highest_deliverable) missing.removeLowerThan(highest_deliverable + 1); retransmit(missing, target); } } } else if(!xmit_task_map.isEmpty()) xmit_task_map.remove(target); // no current gaps for target } // sender: only send the *highest sent* message if HA < HS and HA/HS didn't change from the prev run for(SenderEntry val: send_table.values()) { Table win=val != null? val.msgs : null; if(win != null) { long highest_acked=win.getHighestDelivered(); // highest delivered == highest ack (sender win) long highest_sent=win.getHighestReceived(); // we use table as a *sender* win, so it's highest *sent*... if(highest_acked < highest_sent && val.watermark[0] == highest_acked && val.watermark[1] == highest_sent) { // highest acked and sent hasn't moved up - let's resend the HS Message highest_sent_msg=win.get(highest_sent); if(highest_sent_msg != null) retransmit(highest_sent_msg); } else val.watermark(highest_acked, highest_sent); } } // close idle connections if(conn_expiry_timeout > 0) closeIdleConnections(); if(conn_close_timeout > 0) removeExpiredConnections(); } @ManagedOperation(description="Sends ACKs immediately for entries which are marked as pending (ACK hasn't been sent yet)") public void sendPendingAcks() { for(Map.Entry entry: recv_table.entrySet()) { Address target=entry.getKey(); // target to send retransmit requests to ReceiverEntry val=entry.getValue(); Table win=val != null? val.msgs : null; // receiver: send ack for received messages if needed if(win != null && val.needToSendAck())// sendAck() resets send_ack to false sendAck(target, win.getHighestDeliverable(), val.connId()); } } protected void update(Entry entry, int num_received) { if(conn_expiry_timeout > 0) entry.update(); if(entry.state() == State.CLOSING) entry.state(State.OPEN); num_msgs_received+=num_received; } /** Compares 2 timestamps, handles numeric overflow */ protected static int compare(int ts1, int ts2) { int diff=ts1 - ts2; return Integer.compare(diff, 0); } @SafeVarargs protected static int accumulate(ToIntFunction> func, Collection ... entries) { return Stream.of(entries).flatMap(Collection::stream) .map(entry -> entry.msgs).filter(Objects::nonNull) .mapToInt(func).sum(); } protected enum State {OPEN, CLOSING, CLOSED} protected abstract class Entry { protected final Table msgs; // stores sent or received messages protected final short conn_id; protected final AtomicLong timestamp=new AtomicLong(0); // ns protected volatile State state=State.OPEN; protected Entry(short conn_id, Table msgs) { this.conn_id=conn_id; this.msgs=msgs; update(); } short connId() {return conn_id;} void update() {timestamp.set(getTimestamp());} State state() {return state;} Entry state(State state) {if(this.state != state) {this.state=state; update();} return this;} /** Returns the age of the entry in ms */ long age() {return TimeUnit.MILLISECONDS.convert(getTimestamp() - timestamp.longValue(), TimeUnit.NANOSECONDS);} } protected final class SenderEntry extends Entry { final AtomicLong sent_msgs_seqno=new AtomicLong(DEFAULT_FIRST_SEQNO); // seqno for msgs sent by us final long[] watermark={0,0}; // the highest acked and highest sent seqno int last_timestamp; // to prevent out-of-order ACKs from a receiver public SenderEntry(short send_conn_id) { super(send_conn_id, new Table<>(xmit_table_num_rows, xmit_table_msgs_per_row, 0, xmit_table_resize_factor, xmit_table_max_compaction_time)); } long[] watermark() {return watermark;} SenderEntry watermark(long ha, long hs) {watermark[0]=ha; watermark[1]=hs; return this;} /** Updates last_timestamp. Returns true of the update was in order (ts > last_timestamp) */ private synchronized boolean updateLastTimestamp(int ts) { if(last_timestamp == 0) { last_timestamp=ts; return true; } boolean success=compare(ts, last_timestamp) > 0; // ts has to be > last_timestamp if(success) last_timestamp=ts; return success; } public String toString() { StringBuilder sb=new StringBuilder(); if(msgs != null) sb.append(msgs).append(", "); sb.append("send_conn_id=" + conn_id).append(" (" + age()/1000 + " secs old) - " + state); if(last_timestamp != 0) sb.append(", last-ts: ").append(last_timestamp); return sb.toString(); } } protected final class ReceiverEntry extends Entry { private final AtomicBoolean send_ack=new AtomicBoolean(); public ReceiverEntry(Table received_msgs, short recv_conn_id) { super(recv_conn_id, received_msgs); } ReceiverEntry sendAck() {send_ack.compareAndSet(false, true); return this;} boolean needToSendAck() {return send_ack.compareAndSet(true, false);} public String toString() { StringBuilder sb=new StringBuilder(); if(msgs != null) sb.append(msgs).append(", "); sb.append("recv_conn_id=" + conn_id).append(" (" + age() / 1000 + " secs old) - " + state); if(send_ack.get()) sb.append(" [ack pending]"); return sb.toString(); } } /** * Retransmitter task which periodically (every xmit_interval ms): *
    *
  • If any of the receiver windows have the ack flag set, clears the flag and sends an ack for the * highest delivered seqno to the sender
  • *
  • Checks all receiver windows for missing messages and asks senders for retransmission
  • *
  • For all sender windows, checks if highest acked (HA) < highest sent (HS). If not, and HA/HS is the same * as on the last retransmission run, send the highest sent message again
  • *
*/ protected class RetransmitTask implements Runnable { public void run() { triggerXmit(); } public String toString() { return UNICAST3.class.getSimpleName() + ": RetransmitTask (interval=" + xmit_interval + " ms)"; } } /** * Used to queue messages until a {@link ReceiverEntry} has been created. Queued messages are then removed from * the cache and added to the ReceiverEntry */ protected class MessageCache { private final Map> map=new ConcurrentHashMap<>(); private volatile boolean is_empty=true; protected MessageCache cache(Address sender, Message msg) { List list=map.computeIfAbsent(sender, addr -> new ArrayList<>()); list.add(msg); is_empty=false; return this; } protected List drain(Address sender) { List list=map.remove(sender); if(map.isEmpty()) is_empty=true; return list; } protected MessageCache clear() { map.clear(); is_empty=true; return this; } /** Returns a count of all messages */ protected int size() { return map.values().stream().mapToInt(Collection::size).sum(); } protected boolean isEmpty() { return is_empty; } public String toString() { return String.format("%d message(s)", size()); } } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy