Switch to side-by-side view

--- a
+++ b/examples/SimpleRelayCongestion/description.txt
@@ -0,0 +1,108 @@
+# This topology (SimpleRelayWithCongestion) is designed for testing simple 
+# network congestion on a InteriorRouter. You can change its behaviour by 
+# defining multiple variables for EFCP, RMT and FA. Several implications needs 
+# to be taken into account when defining these variables and their mutual relations.
+# For Flow Allocator it is mainly the createRequestTimeout. It specifies the 
+# upper limit in which it expects response for createFlow request. So if you want 
+# to avoid re-sending createFlow request, set it high enough to accomodate for 
+# recursive Flow creation. It depends heavily on chosen topology.
+# Default value is 10s. You can change it in .ini.
+#
+# Example:
+# **.fa.createRequestTimeout = 12s
+#
+# For EFCP it is
+# initialSenderCredit (default 10)
+# rcvCredit (default 10)
+# closedWindowQLen (default 4)
+# mpl (default 50s)
+# rtt (default 2s)
+# initialSenderCredit 	specifies the initial credit on the sender direction. 
+#					   	Setting this variable high (might) cause the 
+#						InteriorRouter get congested. Setting it too low will on 
+#						the other hand prevent the initial overload. After first 
+#						exchange of FlowControlPDU, the initialSenderCredit looses 
+#						its importance.
+# rcvCredit 			is the value that is send (in form of RcvRightWindowEdge) 
+#						in mentioned FlowControlPDU.
+# closedWindowQLen 		takes place when SndRightWindowEdge permits from sending 
+#						another PDU, so DTP starts to put them on closedWindowQ. 
+#						When lenght of this queue reaches closedWindowQLen, 
+#						DTP signals Push-Back to upper flow.
+# mpl 					is maximum PDU lifetime and is used for computation few 
+#						inner timers. This variable is fixed throughout simulation run.
+# rtt 					is initial value for round-trip time. RTT is then gradually 
+#						updated through RTTEstimator policy. Set it high enough, 
+#						so the first DataTransferPDU - Ack exchange have chance 
+#						to updated it before retransmission timer expiration.
+#
+# In RMT you can set defaultThreshQLength, defaultMaxQLength and maxQPolicyName.
+# defaultThreshQLength 	sets initial threshold length of dynamically initiated RMT 
+#						queues (10 by default).
+# defaultMaxQLength 	sets initial maximum length of dynamically initiated RMT 
+#						queues (20 by default, defaultMaxQLength >= defaultThreshQLength). 
+# maxQPolicyName 		specifies the RMT MaxQueue policy to be used in the scenario. 
+#						This policy is invoked by a queue each time number of 
+#						queued PDUs exceeds the defaultThreshQLength. The sample 
+#						MaxQueue policy used for invocation of congestion notifications 
+#						is called "UpstreamNotifier".
+#
+# CongestionPing:
+# We are trying to congest InteriorRouter so the connection between InteriorRouter 
+# and Host2 has bigger latency, queues from InteriorRouter to Host2 are shorter, etc.
+#
+# Used AE: AEPing - behave in same sense as ICMP Echo Request/reply
+# **.host1.applicationProcess1.applicationEntity.iae.dstApName = "App2"
+# **.host1.applicationProcess1.applicationEntity.iae.dstAeName = "Ping"
+# **.host1.applicationProcess1.applicationEntity.iae.startAt = 10s -> start of Application entity (not start of sending PINGs)
+# **.host1.applicationProcess1.applicationEntity.iae.pingAt =  60s -> AE starts sending PINGs
+# **.host1.applicationProcess1.applicationEntity.iae.rate = 50 -> number of PINGs that will be sent
+# **.host1.applicationProcess1.applicationEntity.iae.stopAt = 250s -> time of deallocation
+# **.interiorRouter.relayIpc.relayAndMux.defaultMaxQLength = 5 -> shortened output queue towards Host2
+# **.interiorRouter.relayIpc.relayAndMux.defaultThreshQLength = 3 -> lowered threshold
+# **.efcp.rtt = 25s -> high enough
+# **.host1.ipcProcess1.efcp.efcp.initialSenderCredit = 50 -> Host1 can send up to 50 PDUs before getting Ack or FlowControl update
+# **.interiorRouter.ipcProcess1.efcp.efcp.initialSenderCredit = 3 -> But IPC towards Host2 on InteriorRouter in the lower DIF can send only up to 3 PDUs before getting Ack or FlowControl update
+# **.interiorRouter.ipcProcess1.efcp.efcp.maxClosedWinQueLen = 4 -> after reaching 4 PDUs on closedWindowQ this IPC emits Push-Back
+# **.host2.ipcProcess0.efcp.efcp.rcvCredit = 3 -> thanks to this, the sender credit on Interior router in ipcProcess1 stays same as initialSenderCredit even after FlowControl update.
+# **.relayAndMux.maxQPolicyName = "UpstreamNotifier" -> Name of policy.
+#
+# Important Events:
+# t=10s - created connection between Host1.ipcp1 - Host1.ipcp0
+# t=10.000Xs - created connection between Host1.ipcp0 - interiorRouter.ipcp0, interiorRouter.ipcp0 - interiorRouter.relayIpc, interiorRouter.relayIpc - interiorRouter.ipcp1.
+# t=15s - created connection between interiorRouter.ipcp1 - Host2.ipcp0.
+# t=25s - created connection Host2.ipcp0 - Host2.ipcp1, Host2.ipcp1 - Host2.irm
+# t=30s - created connection Host1.ipcp1 - Host1.irm
+# t=60s - start of sending PINGs
+# t=62s - interiorRouter.ipcp1 - senderCredit gets depleted (see SndRightWindowEdge -5 and NextSeqNumToSend -6)
+# t=63s - interiorRouter.ipcp1 - first PDU is put on closedWindowQ
+# t=66s - interiorRouter.ipcp1 - closedWindowQ is full - initiate PushBack (block upper flow),
+#       -               .relayIpc - RMT shuts down port towards ipcp1 and it starts to fill
+# t=70s - interiorRouter.ipcp1 - Ack is received -> there is space in closedWindowQ -> RMT port is unblocked and 1 PDU is released.
+# t=70s - interiorRouter.ipcp1 - closedWindowQ is full - initiate PushBack (block upper flow) (this happens several times
+# t=95s - interiorRouter.relayIpc - RMT port is full - SlowDown mechanism is invoked
+#       -                         - RIBd sends CDAP message to Host1.ipcp1 to "SlowDown"
+# t=95s - Host1.ipcp1 - RIBd receives CDAP message to "SlowDown" and ECNSlowDownPolicy in DTCP is initiated.
+#
+#
+# CongestionStream
+# Used AE: AEStream - Sends messages to the other side. Unlike AEPing, it does not send response back.
+# Important Events:
+# t=10s - created connection between Host1.ipcp1 - Host1.ipcp0
+# t=10.000Xs - created connection between Host1.ipcp0 - interiorRouter.ipcp0, interiorRouter.ipcp0 - interiorRouter.relayIpc, interiorRouter.relayIpc - interiorRouter.ipcp1.
+# t=15s - created connection between interiorRouter.ipcp1 - Host2.ipcp0.
+# t=25s - created connection Host2.ipcp0 - Host2.ipcp1, Host2.ipcp1 - Host2.irm
+# t=30s - created connection Host1.ipcp1 - Host1.irm
+#
+# t=60s - start of sending PINGs
+# t=62s - interiorRouter.ipcp1 - senderCredit gets depleted (see SndRightWindowEdge -5 and NextSeqNumToSend -6)
+# t=63s - interiorRouter.ipcp1 - first PDU is put on closedWindowQ
+# t=66s - interiorRouter.ipcp1 - closedWindowQ is full - initiate PushBack (block upper flow),
+#       -               .relayIpc - RMT shuts down port towards ipcp1 and it starts to fill
+# t=70s - interiorRouter.ipcp1 - Ack is received -> there is space in closedWindowQ -> RMT port is unblocked and 1 PDU is released.
+# t=70s - interiorRouter.ipcp1 - closedWindowQ is full - initiate PushBack (block upper flow),
+# t=84s - interiorRouter.relayIpc - RMT port is full - SlowDown mechanism is invoked
+#       -                         - RIBd sends CDAP message to Host1.ipcp1 to "SlowDown"
+# t=84s - Host1.ipcp1 - RIBd receives CDAP message to "SlowDown" and ECNSlowDownPolicy in DTCP is initiated.  
+
+