2 research outputs found
Bias in Internet Measurement Platforms
Network operators and researchers frequently use Internet measurement
platforms (IMPs), such as RIPE Atlas, RIPE RIS, or RouteViews for, e.g.,
monitoring network performance, detecting routing events, topology discovery,
or route optimization. To interpret the results of their measurements and avoid
pitfalls or wrong generalizations, users must understand a platform's
limitations. To this end, this paper studies an important limitation of IMPs,
the \textit{bias}, which exists due to the non-uniform deployment of the
vantage points. Specifically, we introduce a generic framework to
systematically and comprehensively quantify the multi-dimensional (e.g., across
location, topology, network types, etc.) biases of IMPs. Using the framework
and open datasets, we perform a detailed analysis of biases in IMPs that
confirms well-known (to the domain experts) biases and sheds light on
less-known or unexplored biases. To facilitate IMP users to obtain awareness of
and explore bias in their measurements, as well as further research and
analyses (e.g., methods for mitigating bias), we publicly share our code and
data, and provide online tools (API, Web app, etc.) that calculate and
visualize the bias in measurement setups
Filtering the noise to reveal inter-domain lies
On the Internet, routers of Autonomous Systems (ASes) have to determine their preferred inter-domain route, i.e. control path (CP), for each IP prefix. The traffic is then forwarded AS after AS, following a data path (DP) that should match the CP for the same prefix. The underlying implicit trust that ASes advertise the paths they use for packet forwarding may be misplaced. Network operators may tweak CPs and DPs to carry out inter-domain lies that are visible when the two paths differ. Lies can be either unintended, due to misconfigurations or technical limitations, or deliberate, e.g. for economical gain. While lies globally mitigate the ability to troubleshoot and understand the root cause of connectivity issues, detecting them is not a trivial task as the ground data is noisy.In this paper, we propose a modular framework to measure and correctly quantify the discrepancies between CPs and DPs. We define several rules to overcome specific sources of noise inducing mismatches (MMs), e.g., incomplete traces, sibling ASes, IXPs or third-party addresses in general. We leverage the Peering testbed to conduct a measurement campaign at a scale never achieved before, and conclude that, while the upper bound of lies is significant, the lower bound is not negligible. This suggests that the noise interfering with collected traces is not the sole culprit for the MMs between CPs and DPs