The Trade-Off Between Power Consumption and Latency in Computer Networks
As the power consumed by computer networks is nearly independent from the load, networks consume much more power than necessary when the load is low. The authors analyze the influence of disabling network components on power consumption and latency of data transfers. They define two power consumption models and compare the power consumption necessary to achieve given upper limits of latency. Their results show that a trade-off between power consumption and latency exists; it is e.g., possible to conserve 39% of power when accepting a 20% latency increase in the nobel-germany network. They conclude that it is possible for networks to adapt to changing demands to conserve power while the latencies increase only slightly.