Fabrizio Grandoni
University of Lugano
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Fabrizio Grandoni.
Journal of the ACM | 2009
Fedor V. Fomin; Fabrizio Grandoni; Dieter Kratsch
For more than 40 years, Branch & Reduce exponential-time backtracking algorithms have been among the most common tools used for finding exact solutions of NP-hard problems. Despite that, the way to analyze such recursive algorithms is still far from producing tight worst-case running time bounds. Motivated by this, we use an approach, that we call “Measure & Conquer””, as an attempt to step beyond such limitations. The approach is based on the careful design of a nonstandard measure of the subproblem size; this measure is then used to lower bound the progress made by the algorithm at each branching step. The idea is that a smarter measure may capture behaviors of the algorithm that a standard measure might not be able to exploit, and hence lead to a significantly better worst-case time analysis. In order to show the potentialities of Measure & Conquer, we consider two well-studied NP-hard problems: minimum dominating set and maximum independent set. For the first problem, we consider the current best algorithm, and prove (thanks to a better measure) a much tighter running time bound for it. For the second problem, we describe a new, simple algorithm, and show that its running time is competitive with the current best time bounds, achieved with far more complicated algorithms (and standard analysis). Our examples show that a good choice of the measure, made in the very first stages of exact algorithms design, can have a tremendous impact on the running time bounds achievable.
international colloquium on automata languages and programming | 2005
Fedor V. Fomin; Fabrizio Grandoni; Dieter Kratsch
Davis-Putnam-style exponential-time backtracking algorithms are the most common algorithms used for finding exact solutions of NP-hard problems. The analysis of such recursive algorithms is based on the bounded search tree technique: a measure of the size of the subproblems is defined; this measure is used to lower bound the progress made by the algorithm at each branching step. For the last 30 years the research on exact algorithms has been mainly focused on the design of more and more sophisticated algorithms. However, measures used in the analysis of backtracking algorithms are usually very simple. In this paper we stress that a more careful choice of the measure can lead to significantly better worst case time analysis. As an example, we consider the minimum dominating set problem. The currently fastest algorithm for this problem has running time O(20.850n) on n-nodes graphs. By measuring the progress of the (same) algorithm in a different way, we refine the time bound to O(20.598 n). A good choice of the measure can provide such a (surprisingly big) improvement; this suggests that the running time of many other exponential-time recursive algorithms is largely overestimated because of a “bad” choice of the measure.
Journal of the ACM | 2013
Jaroslaw Byrka; Fabrizio Grandoni; Thomas Rothvoss; Laura Sanità
The Steiner tree problem is one of the most fundamental NP-hard problems: given a weighted undirected graph and a subset of terminal nodes, find a minimum-cost tree spanning the terminals. In a sequence of papers, the approximation ratio for this problem was improved from 2 to 1.55 [Robins and Zelikovsky 2005]. All these algorithms are purely combinatorial. A long-standing open problem is whether there is an LP relaxation of Steiner tree with integrality gap smaller than 2 [Rajagopalan and Vazirani 1999]. In this article we present an LP-based approximation algorithm for Steiner tree with an improved approximation factor. Our algorithm is based on a, seemingly novel, iterative randomized rounding technique. We consider an LP relaxation of the problem, which is based on the notion of directed components. We sample one component with probability proportional to the value of the associated variable in a fractional solution: the sampled component is contracted and the LP is updated consequently. We iterate this process until all terminals are connected. Our algorithm delivers a solution of cost at most ln(4) + ϵ < 1.39 times the cost of an optimal Steiner tree. The algorithm can be derandomized using the method of limited independence. As a by-product of our analysis, we show that the integrality gap of our LP is at most 1.55, hence answering the mentioned open question.
Algorithmica | 2008
Fedor V. Fomin; Fabrizio Grandoni; Dieter Kratsch
Abstract In the connected dominating set problem we are given an n-node undirected graph, and we are asked to find a minimum cardinality connected subset S of nodes such that each node not in S is adjacent to some node in S. This problem is also equivalent to finding a spanning tree with maximum number of leaves. Despite its relevance in applications, the best known exact algorithm for the problem is the trivial Ω(2n) algorithm that enumerates all the subsets of nodes. This is not the case for the general (unconnected) version of the problem, for which much faster algorithms are available. Such a difference is not surprising, since connectivity is a global property, and non-local problems are typically much harder to solve exactly. In this paper we break the 2n barrier, by presenting a simple O(1.9407n) algorithm for the connected dominating set problem. The algorithm makes use of new domination rules, and its analysis is based on the Measure and Conquer technique.
ACM Transactions on Algorithms | 2008
Fedor V. Fomin; Fabrizio Grandoni; Artem V. Pyatkin; Alexey A. Stepanov
We provide an algorithm listing all minimal dominating sets of a graph on <i>n</i> vertices in time <i>O</i>(1.7159<sup><i>n</i></sup>). This result can be seen as an algorithmic proof of the fact that the number of minimal dominating sets in a graph on <i>n</i> vertices is at most 1.7159<sup><i>n</i></sup>, thus improving on the trivial <i>O</i>(2<sup><i>n</i></sup>/&sqrt;<i>n</i>) bound. Our result makes use of the measure-and-conquer technique which was recently developed in the area of exact algorithms. Based on this result, we derive an <i>O</i>(2.8718<sup><i>n</i></sup>) algorithm for the domatic number problem.
Journal of Discrete Algorithms | 2006
Fabrizio Grandoni
The currently (asymptotically) fastest algorithm for minimum dominating set on graphs of n nodes is the trivial @W(2^n) algorithm which enumerates and checks all the subsets of nodes. In this paper we present a simple algorithm which solves this problem in O(1.81^n) time.
Theoretical Computer Science | 2004
Friedrich Eisenbrand; Fabrizio Grandoni
We provide simple, faster algorithms for the detection of cliques and dominating sets of fixed order. Our algorithms are based on reductions to rectangular matrix multiplication. We also describe an improved algorithm for diamonds detection.
Information Processing Letters | 2005
L. Sunil Chandran; Fabrizio Grandoni
Memorization is a technique which allows to speed up exponential recursive algorithms at the cost of an exponential space complexity. This technique already leads to the currently fastest algorithm for fixed-parameter vertex cover, whose time complexity is O(1.2832^kk^1^.^5+kn), where n is the number of nodes and k is the size of the vertex cover. Via a refined use of memorization, we obtain an O(1.2759^kk^1^.^5+kn) algorithm for the same problem. We moreover show how to further reduce the complexity to O(1.2745^kk^4+kn).
Journal of Computer and System Sciences | 2010
Friedrich Eisenbrand; Fabrizio Grandoni; Thomas Rothvoí; Guido Schäfer
We present a simple randomized algorithmic framework for connected facility location problems. The basic idea is as follows: We run a black-box approximation algorithm for the unconnected facility location problem, randomly sample the clients, and open the facilities serving sampled clients in the approximate solution. Via a novel analytical tool, which we term core detouring, we show that this approach significantly improves over the previously best known approximation ratios for several NP-hard network design problems. For example, we reduce the approximation ratio for the connected facility location problem from 8.55 to 4.00 and for the single-sink rent-or-buy problem from 3.55 to 2.92. The mentioned results can be derandomized at the expense of a slightly worse approximation ratio. The versatility of our framework is demonstrated by devising improved approximation algorithms also for other related problems.
symposium on discrete algorithms | 2007
Irene Finocchi; Fabrizio Grandoni; Giuseppe F. Italiano
We investigate the problem of computing in a reliable fashion in the presence of faults that may arbitrarily corrupt memory locations. In this framework, we focus on the design of resilient data structures, i.e., data structures that, despite the corruption of some memory values during their lifetime, are nevertheless able to operate correctly (at least) on the set of uncorrupted values. In particular, we present resilient search trees which achieve optimal time and space bounds while tolerating up to <i>O</i>(√log <i>n</i>) memory faults, where <i>n</i> is the current number of items in the search tree. In more detail, our resilient search trees are able to insert, delete and search for a key in <i>O</i>(log <i>n</i> + <i>Δ</i><sup>2</sup>) amortized time, where Δ is an upper bound on the total number of faults. The space required is <i>O</i>(<i>n</i> + <i>Δ</i>).