Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Discover millions of ebooks, audiobooks, and so much more with a free trial

From $11.99/month after trial. Cancel anytime.

Concepts of Combinatorial Optimization
Concepts of Combinatorial Optimization
Concepts of Combinatorial Optimization
Ebook673 pages7 hours

Concepts of Combinatorial Optimization

Rating: 0 out of 5 stars

()

Read preview

About this ebook

Combinatorial optimization is a multidisciplinary scientific area, lying in the interface of three major scientific domains: mathematics, theoretical computer science and management.  The three volumes of the Combinatorial Optimization series aim to cover a wide range  of topics in this area. These topics also deal with fundamental notions and approaches as with several classical applications of combinatorial optimization.

Concepts of Combinatorial Optimization, is divided into three parts:
- On the complexity of combinatorial optimization problems, presenting basics about worst-case and randomized complexity;
- Classical solution methods, presenting the two most-known methods for solving hard combinatorial optimization problems, that are Branch-and-Bound and Dynamic Programming;
- Elements from mathematical programming, presenting fundamentals from mathematical programming based methods that are in the heart of Operations Research since the origins of this field.
LanguageEnglish
PublisherWiley
Release dateAug 8, 2014
ISBN9781119015079
Concepts of Combinatorial Optimization

Related to Concepts of Combinatorial Optimization

Related ebooks

Mathematics For You

View More

Related articles

Reviews for Concepts of Combinatorial Optimization

Rating: 0 out of 5 stars
0 ratings

0 ratings0 reviews

What did you think?

Tap to rate

Review must be at least 10 words

    Book preview

    Concepts of Combinatorial Optimization - Vangelis Th. Paschos

    Preface

    What is combinatorial optimization? There are, in my opinion, as many definitions as there are researchers in this domain, each one as valid as the other. For me, it is above all the art of understanding a real, natural problem, and being able to transform it into a mathematical model. It is the art of studying this model in order to extract its structural properties and the characteristics of the solutions of the modeled problem. It is the art of exploiting these characteristics in order to determine algorithms that calculate the solutions but also to show the limits in economy and efficiency of these algorithms. Lastly, it is the art of enriching or abstracting existing models in order to increase their strength, portability, and ability to describe mathematically (and computationally) other problems, which may or may not be similar to the problems that inspired the initial models.

    Seen in this light, we can easily understand why combinatorial optimization is at the heart of the junction of scientific disciplines as rich and different as theoretical computer science, pure and applied, discrete and continuous mathematics, mathematical economics, and quantitative management. It is inspired by these, and enriches them all.

    This revised and updated second edition of Concepts of Combinatorial Optimization, is the first volume in a series of books entitled Combinatorial Optimization. It tries, along with the other volumes in the set, to embody the idea of combinatorial optimization. The subjects of this volume cover themes that are considered to constitute the hard core of combinatorial optimization. The book is divided into three parts:

    – Part I: Complexity of Combinatorial Optimization Problems;

    – Part II: Classical Solution Methods;

    – Part III: Elements from Mathematical Programming.

    In the first part, Chapter 1 introduces the fundamentals of the theory of (deterministic) complexity and of algorithm analysis. In Chapter 2, the context changes and we consider algorithms that make decisions by tossing a coin. At each stage of the resolution of a problem, several alternatives have to be considered, each one occurring with a certain probability. This is the context of probabilistic (or randomized) algorithms, which is described in this chapter.

    In the second part some methods are introduced that make up the great classics of combinatorial optimization: branch-and-bound and dynamic programming. The former is perhaps the most well known and the most popular when we try to find an optimal solution to a difficult combinatorial optimization problem. Chapter 3 gives a thorough overview of this method as well as of some of the most well-known tree search methods based upon branch-and-bound. What can we say about dynamic programming, presented in Chapter 4? It has considerable reach and scope, and very many optimization problems have optimal solution algorithms that use it as their central method.

    The third part is centered around mathematical programming, considered to be the heart of combinatorial optimization and operational research. In Chapter 5, a large number of linear models and an equally large number of combinatorial optimization problems are set out and discussed. In Chapter 6, the main simplex algorithms for linear programming, such as the primal simplex algorithm, the dual simplex algorithm, and the primal–dual simplex algorithm are introduced. Chapter 7 introduces some classical linear programming methods, while Chapter 8 introduces quadratic integer optimization methods. Chapter 9 describes a series of resolution methods currently widely in use, namely column generation. Chapter 10 focuses on polyhedral methods, almost 60 years old but still relevant to combinatorial optimization research. Lastly, Chapter 11 introduces a more contemporary, but extremely interesting, subject, namely constraint programming.

    This book is intended for novice researchers, or even Master’s students, as much as for senior researchers. Master’s students will probably need a little basic knowledge of graph theory and mathematical (especially linear) programming to be able to read the book comfortably, even though the authors have been careful to give definitions of all the concepts they use in their chapters. In any case, to improve their knowledge of graph theory, readers are invited to consult a great, flagship book from one of our gurus, Claude Berge: Graphs and Hypergraphs, North Holland, 1973. For linear programming, there is a multitude of good books that the reader could consult, for example V. Chvátal, Linear Programming, W.H. Freeman, 1983, or M. Minoux, Programmation mathématique: théorie et algorithmes, Dunod, 1983.

    In this revised second edition I would like to thank again the authors who, despite their many responsibilities and commitments (which is the lot of any university academic), agreed to participate in the book by writing chapters in their areas of expertise and, at the same time, to take part in a very tricky exercise: writing chapters that are both educational and high-level science at the same time.

    The second edition of this book could never have come into being without the original proposal of Jean-Charles Pomerol, President of the scientific board at ISTE, and Sami Ménascé and Raphaël Ménascé, the President and Vice-President of ISTE, respectively. I once again give my warmest thanks to them for their insistence and encouragement.

    Vangelis Th. PASCHOS

    June 2014

    PART I

    Complexity of Combinatorial Optimization Problems

    Chapter 1

    Basic Concepts in Algorithms and Complexity Theory

    Vangelis Th. PASCHOS.

    1.1. Algorithmic complexity

    In algorithmic theory, a problem is a general question to which we wish to find an answer. This question usually has parameters or variables the values of which have yet to be determined. A problem is posed by giving a list of these parameters as well as the properties to which the answer must conform. An instance of a problem is obtained by giving explicit values to each of the parameters of the instanced problem.

    An algorithm is a sequence of elementary operations (variable affectation, tests, forks, etc.) that, when given an instance of a problem as input, gives the solution of this problem as output after execution of the final operation.

    The two most important parameters for measuring the quality of an algorithm are: its execution time and the memory space that it uses. The first parameter is expressed in terms of the number of instructions necessary to run the algorithm. The use of the number of instructions as a unit of time is justified by the fact that the same program will use the same number of instructions on two different machines but the time taken will vary, depending on the respective speeds of the machines. We generally consider that an instruction equates to an elementary operation, for example an assignment, a test, an addition, a multiplication, a trace, etc. What we call the complexity in time or simply the complexity of an algorithm gives us an indication of the time it will take to solve a problem of a given size. In reality this is a function that associates an order of magnitude¹ of the number of instructions necessary for the solution of a given problem with the size of an instance of that problem. The second parameter corresponds to the number of memory units used by the algorithm to solve a problem. The complexity in space is a function that associates an order of magnitude of the number of memory units used for the operations necessary for the solution of a given problem with the size of an instance of that problem.

    There are several sets of hypotheses concerning the standard configuration that we use as a basis for measuring the complexity of an algorithm. The most commonly used framework is the one known as worst-case. Here, the complexity of an algorithm is the number of operations carried out on the instance that represents the worst configuration, amongst those of a fixed size, for its execution; this is the framework used in most of this book. However, it is not the only framework for analyzing the complexity of an algorithm. Another framework often used is average analysis. This kind of analysis consists of finding, for a fixed size (of the instance) n, the average execution time of an algorithm on all the instances of size n; we assume that for this analysis the probability of each instance occurring follows a specific distribution pattern. More often than not, this distribution pattern is considered to be uniform. There are three main reasons for the worst-case analysis being used more often than the average analysis. The first is psychological: the worst-case result tells us for certain that the algorithm being analyzed can never have a level of complexity higher than that shown by this analysis; in other words, the result we have obtained gives us an upper bound on the complexity of our algorithm. The second reason is mathematical: results from a worst-case analysis are often easier to obtain than those from an average analysis, which very often requires mathematical tools and more complex analysis. The third reason is analysis portability: the validity of an average analysis is limited by the assumptions made about the distribution pattern of the instances; if the assumptions change, then the original analysis is no longer valid.

    1.2. Problem complexity

    The definition of the complexity of an algorithm can be easily transposed to problems. Informally, the complexity of a problem is equal to the complexity of the best algorithm that solves it (this definition is valid independently of which framework we use).

    Let us take a size n and a function f (n). Thus:

    TIMEf (n) is the class of problems for which the complexity (in time) of an instance of size n is in O(f (n)).

    SPACEf (n) is the class of problems that can be solved, for an instance of size n, by using a memory space of O(f (n)).

    We can now specify the following general classes of complexity:

    P is the class of all the problems that can be solved in a time that is a polynomial function of their instance size, that is .

    EXPTIME is the class of problems that can be solved in a time that is an exponential function of their instance size, that is EXPTIME = .

    PSPACE is the class of all the problems that can be solved using a memory space that is a polynomial function of their instance size, that is PSPACE = .

    With respect to the classes that we have just defined, we have the following relations: P PSPACE EXPTIME and P EXPTIME. Knowing whether the inclusions of the first relation are strict or not is still an open problem.

    Almost all combinatorial optimization problems can be classified, from an algorithmic complexity point of view, into two large categories. Polynomial problems can be solved optimally by algorithms of polynomial complexity, that is in O(nk), where k is a constant independent of n (this is the class P that we have already defined). Non-polynomial problems are those for which the best algorithms (those giving an optimum solution) are of a super-polynomial complexity, that is in O(f (n)g (n)), where f and g are increasing functions in n and limn→∞ g (n) = ∞. All these problems contain the class EXPTIME.

    The definition of any algorithmic problem (and even more so in the case of any combinatorial optimization problem) comprises two parts. The first gives the instance of the problem, that is the type of its variables (a graph, a set, a logical expression, etc.). The second part gives the type and properties to which the expected solution must conform. In the complexity theory case, algorithmic problems can be classified into three categories:

    decision problems;

    optimum value calculation problems;

    optimization problems.

    Decision problems are questions concerning the existence, for a given instance, of a configuration such that this configuration itself, or its value, conforms to certain properties. The solution to a decision problem is an answer to the question associated with the problem. In other words, this solution can be:

    – either "yes, such a configuration does exist";

    – or "no, such a configuration does not exist".

    Let us consider as an example the conjunctive normal form satisfiability problem, known in the literature as SAT: "Given a set U of n Boolean variables x1, …, xn and a set of m clauses² C1, …, Cm, is there a model for the expression ϕ = ; i.e. is there an assignment of the values 0 or 1 to the variables such that ϕ = 1?". For an instance ϕ of this problem, if ϕ allows a model then the solution (the correct answer) is yes, otherwise the solution is no.

    Let us now consider the MIN TSP problem, defined as follows: given a complete graph Kn over n vertices for which each edge e E(Kn) has a value d(e) > 0, we are looking for a Hamiltonian cycle H E (a partial closely related graph such that each vertex is of 2 degrees) that minimizes the quantity ∑eH d(e). Let us assume that for this problem we have, as well as the complete graph Kn and the vector , costs on the edges Kn of a constant K and that we are looking not to determine the smallest (in terms of total cost) Hamiltonian cycle, but rather to answer the following question: "Does there exist a Hamiltonian cycle of total distance less than or equal to K?". Here, once more, the solution is either yes if such a cycle exists, or no if it does not.

    For optimum value calculation problems, we are looking to calculate the value of the optimum solution (and not the solution itself).

    In the case of the MIN TSP for example, the optimum associated value calculation problem comes down to calculating the cost of the smallest Hamiltonian cycle, and not the cycle itself.

    Optimization problems, which are naturally of interest to us in this book, are those for which we are looking to establish the best solution amongst those satisfying certain properties given by the very definition of the problem. An optimization problem may be seen as a mathematical program of the form:

    where is the vector describing the solution³, v( ) is the objective function, CI is the problem’s constraint set, set out for the instance I (in other words, CI sets out both the instance and the properties of the solution that we are looking to find for this instance), and opt ∈ {max, min}. An optimum solution of I is a vector ∈ argopt{v( ) : ∈ CI}. The quantity v( ) is known as the objective value or value of the problem. A solution ∈ CI is known as a feasible solution.

    Let us consider the problem MIN WEIGHTED VERTEX COVER⁴. An instance of this problem (given by the information from the incident matrix A, of dimension m × n, of a graph G(V, E) of order n with |E| = m to and a vector , of dimension n of the costs of the edges of V), can be expressed in terms of a linear program in integers as follows:

    where is a vector from 0, 1 of dimension n such that xi = 1 if the vertex Vi V is included in the solution, xi = 0 if it is not included. The block of m constraints expresses the fact that for each edge at least one of these extremes must be included in the solution. The feasible solutions are all the transversals of G and the optimum solution is a transversal of G of minimum total weight, that is a transversal corresponding to a feasible vector consisting of a maximum number of 1.

    The solution to an optimization problem includes an evaluation of the optimum value. Therefore, an optimum value calculation problem can be associated with an optimization problem. Moreover, optimization problems always have a decisional variant as shown in the MIN TSP example above.

    1.3. The classes P, NP and NPO

    Let us consider a decision problem Π. If for any instance I of Π a solution (that is a correct answer to the question that states Π) of I can be found algorithmically in polynomial time, that is in O(|I|k) stages, where |I| is the size of I, then Π is called a polynomial problem and the algorithm that solves it a polynomial algorithm (let us remember that polynomial problems make up the P class).

    For reasons of simplicity, we will assume in what follows that the solution to a decision problem is:

    – either "yes, such a solution exists, and this is it";

    – or "no, such a solution does not exist".

    In other words, if, to solve a problem, we could consult an oracle, it would provide us with an answer of not just a yes or no but also, in the first case, a certificate proving the veracity of the yes. This testimony is simply a solution proposal that the oracle asserts as being the real solution to our problem.

    Let us consider the decision problems for which the validity of the certificate can be verified in polynomial time. These problems form the class NP.

    DEFINITION 1.1.– A decision problem Π is in NP if the validity of all solutions of Π is verifiable in polynomial time.

    For example, the SAT problem belongs to NP. Indeed, given the assignment of the values 0, 1 to the variables of an instance ϕ of this problem, we can, with at most nm applications of the connector , decide whether the proposed assignment is a model for ϕ, that is whether it satisfies all the clauses.

    Therefore, we can easily see that the decisional variant of MIN TSP seen previously also belongs to NP.

    Definition 1.1 can be extended to optimization problems. Let us consider an optimization problem Π and let us assume that each instance I of Π conforms to the three following properties:

    1) The feasibility of a solution can be verified in polynomial time.

    2) The value of a feasible solution can be calculated in polynomial time.

    3) There is at least one feasible solution that can be calculated in polynomial time.

    Thus, Π belongs to the class NPO. In other words, the class NPO is the class of optimization problems for which the decisional variant is in NP. We can therefore define the class PO of optimization problems for which the decisional variant belongs to the class P. In other words, PO is the class of problems that can be optimally solved in polynomial time.

    We note that the class NP has been defined (see definition 1.1) without explicit reference to the optimum solution of its problems, but by reference to the verification of a given solution. Evidently, the condition of belonging to P being stronger than that of belonging to NP (what can be solved can be verified), we have the obvious inclusion of : P NP (Figure 1.1).

    "What is the complexity of the problems in NP \ P?". The best general result on the complexity of the solution of problems from NP is as follows [GAR 79].

    THEOREM 1.1.– For any problem Π ∈ NP, there is a polynomial pΠ such that each instance I of Π can be solved by an algorithm of complexity O(2PΠ(|I|)).

    In fact, theorem 1.1 merely gives an upper limit on the complexity of problems in NP, but no lower limit. The diagram in Figure 1.1 is nothing more than a conjecture, and although almost all researchers in complexity are completely convinced of its veracity, it has still not been proved. The question "Is P equal to or different from NP?" is the biggest open question in computer science and one of the best known in mathematics.

    Figure 1.1. P and NP (under the assumption P NP)

    1.4. Karp and Turing reductions

    As we have seen, problems in NP \ P are considered to be algorithmically more difficult than problems in P. A large number of problems in NP \ P are very strongly bound to each other through the concept of polynomial reduction.

    The principle of reducing a problem n to a problem consists of considering the problem Π as a specific case of , modulo a slight transformation. If this transformation is polynomial, and we know that we can solve in polynomial time, we will also be able to solve Π in polynomial time. Reduction is thus a means of transferring the result of solving one problem to another; in the same way it is a tool for classifying problems according to the level of difficulty of their solution.

    We will start with the classic Karp reduction (for the class NP) [GAR 79, KAR 72]. This links two decision problems by the possibility of their optimum (and simultaneous) solution in polynomial time. In the following, given a problem Π, let be all of its instances (we assume that each instance I ∈ is identifiable in polynomial time in |I|). Let be the subset of for which the solution is yes; is also known as the set of yes-instances (or positive instances) of Π.

    DEFINITION 1.2.– Given two decision problems Π1 and Π2, a Karp reduction (or polynomial transformation) is a function , which can be calculated in polynomial time, such that, given a solution for f(I), we are able to find a solution for I in polynomial time in |I| (the size of the instance I).

    A Karp reduction of a decision problem Π1 to a decision problem Π2 implies the existence of an algorithm for Π1 that uses an algorithm for Π2. Given any instance , the algorithm constructs an instance ; it executes the algorithm , which calculates a solution on I2, then transforms this solution into a solution for Π1 on I1. If is polynomial, then is also polynomial.

    Following on from this, we can state another reduction, known in the literature as the Turing reduction, which is better adapted to optimization problems. In what follows, we define a problem Π as a couple ( , SolΠ), where SolΠ is the set of solutions for Π (we denote by SolΠ(I) the set of solutions for the instance I ∈ ).

    DEFINITION 1.3.– A Turing reduction of a problem Π1 to a problem Π2 is an algorithm that solves Π1 by using (possibly several times) an algorithm for Π2 in such a way that if is polynomial, then is also polynomial.

    The Karp and Turing reductions are transitive: if Π1 is reduced (by one of these two reductions) to Π2 and Π2 is reduced to Π3, then Π1 reduces to Π3. We can therefore see that both reductions preserve membership of the class P in the sense that if Π reduces to and ∈ P, then Π ∈ P.

    For more details on both the Karp and Turing reductions refer to [AUS 99, GAR 79, PAS 04]. In Garey and Johnson [GAR 79] (Chapter 5) there is also a very interesting historical summary of the development of the ideas and terms that have led to the structure of complexity theory as we know it today.

    1.5. NP-completeness

    From the definition of the two reductions in the preceding section, if reduces to Π, then Π can reasonably be considered as at least as difficult as (regarding their solution by polynomial algorithms), in the sense that a polynomial algorithm for Π would have sufficed to solve not only Π itself, but equally . Let us confine ourselves to the Karp reduction. By using it, we can highlight a problem Π* ∈ NP such that any other problem Π ∈ NP reduces to Π* [COO 71, GAR 79, FAG 74]. Such a problem is, as we have mentioned, the most difficult problem of the class NP. Therefore we can show [GAR 79, KAR 72] that there are other problems ∈ NP such that reduces to . In this way we expose a family of problems such that any problem in NP reduces (remembering that the Karp reduction is transitive) to one of its problems. This family has, of course, the following properties:

    – It is made up of the most difficult problems of NP.

    – A polynomial algorithm for at least one of its problems would have been sufficient to solve, in polynomial time, all the other problems of this family (and indeed any problem in NP).

    The problems from this family are NP-complete problems and the class of these problems is called the NP-complete class.

    DEFINITION 1.4.– A decision problem Π is NP-complete if, and only if, it fulfills the following two conditions:

    1) Π ∈ NP ;

    2) ∀Π′ ∈ NP, Π′ reduces to Π by a Karp reduction.

    Of course, a notion of NP-completeness very similar to that of definition 1.4 can also be based on the Turing reduction.

    The following application of definition 1.3 is very often used to show the NP-completeness of a problem. Let Π1 = ( , SolΠ1) and Π2 = be two problems, and let (f, g) be a pair of functions, which can be calculated in polynomial time, where:

    f : is such that for any instance ;

    – is such that for every pair

    .

    Let us assume that there is a polynomial algorithm for the problem Π2. In this case, the algorithm is a (polynomial) Turing reduction.

    A problem that fulfills condition 2 of definition 1.4 (without necessarily checking condition 1) is called NP-hard⁵. It follows that a decision problem Π is NP-complete if and only if it belongs to NP and it is NP-hard.

    With the class NP-complete, we can further refine (Figure 1.2) the world of NP. Of course, if P = NP, the three classes from Figure 1.2 coincide; moreover, under the assumption P NP, the classes P and NP-complete do not intersect.

    Let us denote by NP-intermediate the class NP\ (PNP complete). Informally, this concerns the class of problems of intermediate difficulty, that is problems that are more difficult than those from P but easier than those from NP-complete. More formally, for two complexity classes C and C′ such that C′ ⊂ C, and a reduction preserving the membership of C′, a problem is C-intermediate if it is neither C-complete under , nor belongs to C′. Under the Karp reduction, the class NP-intermediate is not empty [LAD 75].

    Let us note that the idea of NP-completeness goes hand in hand with decision problems. When dealing with optimization problems, the appropriate term, used in the literature, is NP-hard⁶. A problem of NPO is NP-hard if and only if its decisional variant is an NP-complete problem.

    Figure 1.2. P, NP and NP-complete (under the assumption P NP)

    The problem sat was the first problem shown to be NP-complete (the proof of this important result can be found in [COO 71]). The reduction used (often called generic reduction) is based on the theory of recursive languages and Turing machines (see [HOP 79, LEW 81] for more details and depth on the Turing machine concept; also, language-problem correspondence is very well described in [GAR 79, LEW 81]). The general idea of generic reduction, also often called the Cook–Levin technique (or theory), is as follows: For a generic decision (language) problem belonging to NP, we describe, using a normal conjunctive form, the working of a non-deterministic algorithm (Turing machine) that solves (decides) this problem (language).

    The second problem shown to be NP-complete [GAR 79, KAR 72] was the variant of SAT, written 3SAT, where no clause contains more than three literals. The reduction here is from SAT [GAR 79, PAP 81]. More generally, for all k ≥ 3,the kSAT problems (that is the problems defined on normal conjunctive forms where each clause contains no more than k literals) are all NP-complete.

    It must be noted that the problem 2SAT, where all normal conjunctive form clauses contain at most two literals, is polynomial [EVE 76]. It should also be noted that in [KAR 72], where there is a list of the first 21 NP-complete problems, the problem of linear programming in real numbers was mentioned as a probable problem from the class NP-intermediate. It was shown, seven years later ([KHA 79] and also [ASP 80], an English translation of [KHA 79]), that this problem is in P.

    The reference on NP-completeness is the volume by Garey and Johnson [GAR 79]. In the appendix, A list of NP-complete problems, there is a long list of NP-complete problems with several commentaries for each one and for their limited versions. For many years, this list has been regularly updated by Johnson in the Journal of Algorithms review. This update, supplemented by numerous commentaries, appears under the title: The NP-completeness column: an ongoing guide.

    "What is the relationship between optimization and decision for NP-complete problems?". The following theory [AUS 95, CRE 93, PAZ 81] attempts to give an answer.

    THEOREM 1.2.– Let Π be a problem of NPO and let us assume that the decisional version of Π, written Πd, is NP-complete. It follows that a polynomial Turing reduction exists between Πd and Π.

    In other words, the decision versions (such as those we have considered in this chapter) and optimization versions of an NP-complete problem are of equivalent algorithmic difficulty. However, the question of the existence of a problem NPO for which the optimization version is more difficult to solve than its decisional counterpart remains open.

    1.6. Two examples of NP-complete problems

    Given a problem Π, the most conventional way to show its NP-completeness consists of making a Turing or Karp reduction of an NP-complete Π′ problem to Π. In practical terms, the proof of NP-completeness for Π is divided into three stages:

    1) proof of membership of Π to NP;

    2) choice of Π′;

    3) building the functions f and g (see definition 1.3) and showing that they can both be calculated in polynomial time.

    In the following, we show that MIN VERTEX COVER(G(V, E), K) and MAX STABLE(G(V, E), K), the decisional variants of MIN VERTEX COVER and of MAX STABLE⁷, respectively, are NP-complete. These two problems are defined as follows. MIN VERTEX COVER(G(V, E), K): given a graph G and a constant K ≤ |V|, does there exist in G a transversal V′ ⊆ V less than or equal in size to K? MAX STABLE SET(G(V, E), K): given a graph G and a constant K ≤ |V|, does there exist in G a stable set V′ ⊆ V of greater than or equal in size to K?

    1.6.1. MIN VERTEX COVER

    The proof of membership of MIN VERTEX COVER(G(V, E), K) to NP is very simple and so has been omitted here. We will therefore show the completeness of this problem for NP. We will transform an instance ϕ(U, ) from 3SAT, with U = {x1, …, xn} and = {C1, …, Cm}, into an instance (G(V, E), K) of MIN VERTEX COVER.

    This graph is made up of two component sets, joined by edges. The first component is made up of 2n vertices and n edges , which join the vertices in pairs. The second is made up of m vertex-disjoint triangles (that is of m cliques with three vertices). For a clause Ci, we denote the three vertices of the corresponding triangle by ci1, ci2 and ci3. In fact, the first set of components, for which each vertex corresponds to a literal, serves to define the truth values of the solution for 3SAT; the second set of components corresponds to the clauses, and each vertex is associated with a literal of its clause. These triangles are used to verify the satisfaction of the clauses. To finish, we add 3m unifying edges that link each vertex of each triangle-clause to its corresponding literal-vertex. Let us note that exactly three unifying edges go from (the vertices of) each triangle, one per vertex of the triangle. Finally, we state K = n + 2m. It is easy to see that the transformation of ϕ(U, C) to G(V, E) happens in polynomial time in max{m, n} since |V| = 2n + 3m and |E| = n + 6m.

    As an example of the transformation described above, let us consider the instance from 3SAT. The graph G(V, E) for MIN VERTEX COVER is given in Figure 1.3. In this case, K = 12.

    Figure 1.3. The graph associated with the expression

    We will now show that G allows a transversal less than or equal in size to n + 2m if and only if the expression ϕ can be satisfied.

    Let us first show that the condition is necessary. Let us assume that there is a polynomial algorithm that answers the question "Is there in G a transversal V′ ⊆ V of size |V′| ≤ K?", and, if so, returns V′. Let us execute it with K = n + 2m. If the answer from is yes, then the transversal must be of a size equal to n + 2m. In fact, any transversal needs at least n vertices in order to cover the n edges corresponding to the variables of ϕ (one vertex per edge) and 2m vertices to cover the edges of m triangles (two vertices per triangle). As a result, if answers yes, it will have calculated a transversal of exactly n + 2m vertices.

    In the light of the previous observation, given such a transversal V′, we state that xi = 1 if the extremity xi of the edge is taken in V′; if the extremity is included, then we state that = 1, that is xi = 0. We claim that this assignment of the truth values to the variables satisfies ϕ. Indeed, since only one extremity of each edge is taken in V′, only one literal is set to 1 for each variable and, in consequence, the assignment in question is coherent (one, and only one, truth value is assigned to each literal). Moreover, let us consider a triangle Ti of G corresponding to the clause Ci; let us denote its vertices by ci1, ci2 and ci3, and let us assume that the last two belong to V′. Let us also assume that the unifying edge having as an extremity the vertex ci1 is the edge (ci1, lk), lk being one of the literals associated with the variable xk. Since ci1 ∉ V′, lk belongs to it, that is lk = 1, and the existence of the edge (ci1, lk) means that the literal lk belongs to the clause Ci. This is proved by setting lk to 1. By iterating this argument for each clause, the need for the condition is proved. Furthermore, let us note that obtaining the assignment of the truth values to the variables of ϕ is done in polynomial time.

    Let us now show that the condition is also good enough. Given an assignment of truth values satisfying the expression ϕ, let us construct in G a transversal V′ of size n + 2m. To start with, for each variable xi, if xi = 1, then the extremity xi of the edge is put in V′; otherwise, the extremity of the edge is put there. We thereby cover the edges of type , i = 1,…, n, and one unifying edge per triangle. Let Ti (corresponding to the clause Ci), i = 1,…, m, be a triangle and let (lk, ci1) be the unifying edge covered by the setting to 1 of lk. We therefore put the vertices ci2 and ci3 in V′; these vertices cover both the edges of Ti and the two unifying edges having as extremities ci2 and ci3, respectively. By iterating this operation for each triangle, a transversal V′ of size n + 2m is eventually constructed in polynomial time.

    1.6.2. MAX STABLE

    The proof of membership of MAX STABLE(G(V, E), K) is so simple that it is omitted here.

    Let us consider a graph G(V, E) of magnitude n having m edges and let us denote by A its incidence matrix⁸. Let us also consider the expression of MIN VERTEX COVER as a linear program in whole numbers and the transformations that follow:

    We note that the last program in the series is nothing more than the linear program (in whole numbers) of MAX STABLE. Furthermore, this series of equivalents indicates that if a solution vector for MAX STABLE is given, then the vector (that is the vector where we interchange the 1 and the 0 regarding ) is a feasible solution for MIN VERTEX COVER. Furthermore, if contains at least K 1 (that is the size of the stable set is at least equal to K), then the solution vector deduced for MIN VERTEX COVER contains at most n K 1 (that is the size of the transversal is at most equal to n K). Since the function is polynomial, then so is the described transformation.

    1.7. A few words on strong and weak NP-completeness

    Let Π be a problem and I an instance of Π of size |I|. We denote by max(I) the highest number that appears in I. Let us note that max(I) can be exponential in n. An algorithm for Π is known as pseudo-polynomial if it is polynomial in |I| and max(I) (if max(I) is exponential in |I|, then this algorithm is exponential for I).

    DEFINITION 1.5.– An optimization problem is NP-complete in the strong sense (strongly NP-complete) if the problem is NP-complete because of its structure and not because of the size of the

    Enjoying the preview?
    Page 1 of 1