Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Network connectivity under a probabilistic node failure model (2006.13551v1)

Published 24 Jun 2020 in cs.SI and cs.NI

Abstract: Centrality metrics have been widely applied to identify the nodes in a graph whose removal is effective in decomposing the graph into smaller sub-components. The node--removal process is generally used to test network robustness against failures. Most of the available studies assume that the node removal task is always successful. Yet, we argue that this assumption is unrealistic. Indeed, the removal process should take into account also the strength of the targeted node itself, to simulate the failure scenarios in a more effective and realistic fashion. Unlike previous literature, herein a {\em probabilistic node failure model} is proposed, in which nodes may fail with a particular probability, considering two variants, namely: {\em Uniform} (in which the nodes survival-to-failure probability is fixed) and {\em Best Connected} (BC) (where the nodes survival probability is proportional to their degree). To evaluate our method, we consider five popular centrality metrics carrying out an experimental, comparative analysis to evaluate them in terms of {\em effectiveness} and {\em coverage}, on four real-world graphs. By effectiveness and coverage we mean the ability of selecting nodes whose removal decreases graph connectivity the most. Specifically, the graph spectral radius reduction works as a proxy indicator of effectiveness, and the reduction of the largest connected component (LCC) size is a parameter to assess coverage. The metric that caused the biggest drop has been then compared with the Benchmark analysis (i.e, the non-probabilistic degree centrality node removal process) to compare the two approaches. The main finding has been that significant differences emerged through this comparison with a deviation range that varies from 2\% up to 80\% regardless of the dataset used that highlight the existence of a gap between the common practice with a more realistic approach.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Lucia Cavallaro (4 papers)
  2. Stefania Costantini (17 papers)
  3. Pasquale De Meo (31 papers)
  4. Antonio Liotta (27 papers)
  5. Giovanni Stilo (18 papers)
Citations (4)