-
Bug
-
Resolution: Unresolved
-
Major
-
None
-
4.15, 4.15.0, 4.17, 4.17.0
-
Quality / Stability / Reliability
-
False
-
-
None
-
None
-
No
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
Description of problem:
When a pod network load is applied with parallel connections (not just a single connection), running at 4096 byte message sizes.
Version-Release number of selected component (if applicable):
MicroShift 4.15.0-rc5
How reproducible:
Consistently
Steps to Reproduce:
1. Set up MicroShift node (`dnf install microshift`)
2. Start MicroShift service (`systemctl start microshift`)
3. Run super-netperf or similar to test pod to pod network:
super-netperf 2 -H <pod_ip> -l 30 -t TCP_STREAM -- -k rt_latency,p99_latency,throughput,throughput_units,remote_recv_calls,local_send_calls,local_transport_retrans -m 4096 -R 1
Actual results:
We can see relatively large retransmit values, in this example below 86.8 $ cat loss-rt-result-1710248752.csv Type,Driver,Profile,Same node,Host Network,Service,Duration,Parallelism,# of Samples,Message Size,Confidence metric - low,Confidence metric - high,Value TCP Retransmissions,netperf,TCP_STREAM,true,false,false,30,2,5,4096,5712.581064693677,5892.130935306324,86.800000
Expected results:
We do not expect to see high retransmits.
Additional info:
We can also use the network-perf v2 workload:
https://github.com/cloud-bulldozer/e2e-benchmarking/tree/master/workloads/network-perf-v2
With this configuration:
$ cat one.yaml
---
tests:
- TCPStreamMed2:
parallelism: 2
profile: "TCP_STREAM"
duration: 30
samples: 5
messagesize: 4096