Bug #5082
closedndncatchunks low throughput even when retrieving from local content store
0%
Description
Here is the setup: 2 VMs, 50Mbps link (using tc), transferring a 250MB file.
First time - the transfer goes to the producer, as expected. The throughput is around 32Mbps.
Second time - the transfer happens from the local cache (we verified this). However, the throughput is around 40Mbps - I'd expect this to be much higher.
I see some of the issues described here: https://redmine.named-data.net/issues/4362/
Specifically, I am seeing "Send queue length dropped below congestion threshold". See the attached file.
Any suggestions on how we can get higher throughput from the cache?
Files
Updated by Davide Pesavento almost 5 years ago
- Start date deleted (
02/14/2020)
I see many congestion marks... what version of NFD are you using? The marking algorithm was made less aggressive in this commit.
Updated by susmit shannigrahi almost 5 years ago
Davide Pesavento wrote:
I see many congestion marks... what version of NFD are you using? The marking algorithm was made less aggressive in this commit.
0.7.x for ndn-cxx, NFD, and tools.
Updated by susmit shannigrahi almost 5 years ago
Forgot to mention, we tried with --ignore-marks (and with version discovery disabled). Still, the throughput did not increase when fetching from the cache.
Updated by Anonymous almost 5 years ago
Probably CPU bound/general NFD performance issue?
Check the CPU utilization on the VMs. If it's 100%, then that's the issue.
Updated by Anonymous almost 5 years ago
Any suggestions on how we can get higher throughput from the cache?
Disabling logging should help if it's the CPU issue.
I see some of the issues described here: https://redmine.named-data.net/issues/4362/
Not sure what you refer to? The throughput from the cache on my local machine was much higher than 40 Mbps.
Specifically, I am seeing "Send queue length dropped below congestion threshold". See the attached file.
You can safely ignore this message. It's just some trace output about how the congestion marking works, and not a warning.
Updated by susmit shannigrahi almost 5 years ago
ndncatchunks is not utilizing 100% - however, NFD's CPU usage is consistently over 100%. This is probably why the throughput is low.
I played with the init-cwnd (at around 25-30) and now getting higher throughput at around 250-300Mbps from the local cache. This is still low in my opinion but
looks like you were getting similar numbers so maybe that's the upper bound.
Updated by Anonymous almost 5 years ago
ndncatchunks is not utilizing 100% - however, NFD's CPU usage is consistently over 100%. This is probably why the throughput is low.
Yeah, that's what I saw in my earlier measurements as well.
Updated by Davide Pesavento almost 5 years ago
- Subject changed from ndncatchunks low throughput even when retrieving from local content store to ndncatchunks low throughput even when retrieving from local content store
- Status changed from New to Rejected
susmit shannigrahi wrote:
ndncatchunks is not utilizing 100% - however, NFD's CPU usage is consistently over 100%. This is probably why the throughput is low.
Yeah, this sounds like the usual "NFD is slow" problem... not much we can do about it (in the short term).