Date Added: Feb 2012
Empirical studies on link blacklisting show that the delivery rate is very sensitive to the calibration of the blacklisting threshold. If the calibration is too restrictive (the threshold is too high), all neighbors get blacklisted. On the other hand, if the calibration is too loose (the threshold is too low), unreliable links get selected. This paper investigates blacklisting analytically using a model that accounts for the joint effect of the wireless channel (signal strength variance and coherence time) and the network (node density). The model, validated empirically with mote-class hardware, shows that blacklisting does not help if the wireless channel is stable or if the network is relatively sparse.