A FINE-GRAINED APPROACH TO ALGORITHMS AND COMPLEXITY. Virginia Vassilevska Williams Stanford University

Similar documents
A FINE-GRAINED APPROACH TO

Hardness for easy problems. An introduction

6.S078 A FINE-GRAINED APPROACH TO ALGORITHMS AND COMPLEXITY LECTURE 1

Complexity Theory of Polynomial-Time Problems

Complexity Theory of Polynomial-Time Problems

On some fine-grained questions in algorithms and complexity

CSC2420: Algorithm Design, Analysis and Theory Fall 2017

Geometric Problems in Moderate Dimensions

Connections between exponential time and polynomial time problem Lecture Notes for CS294

Matching Triangles and Basing Hardness on an Extremely Popular Conjecture

Simulating Branching Programs with Edit Distance and Friends

Simulating Branching Programs with Edit Distance and Friends Or: A Polylog Shaved is a Lower Bound Made

Deterministic APSP, Orthogonal Vectors, and More:

Algorithm Design and Analysis

CS154, Lecture 15: Cook-Levin Theorem SAT, 3SAT

Complexity Theory of Polynomial-Time Problems

CS/COE

COMP Analysis of Algorithms & Data Structures

NP-Complete Problems. More reductions

Algorithms as Lower Bounds

P is the class of problems for which there are algorithms that solve the problem in time O(n k ) for some constant k.

Strong ETH Breaks With Merlin and Arthur. Or: Short Non-Interactive Proofs of Batch Evaluation

NP-Completeness. Andreas Klappenecker. [based on slides by Prof. Welch]

Notes on Complexity Theory Last updated: October, Lecture 6

Limitations of Algorithm Power

Complete problems for classes in PH, The Polynomial-Time Hierarchy (PH) oracle is like a subroutine, or function in

1 Circuit Complexity. CS 6743 Lecture 15 1 Fall Definitions

Umans Complexity Theory Lectures

Lecture 4: NP and computational intractability

Lecture 8: Complete Problems for Other Complexity Classes

CSE200: Computability and complexity Space Complexity

Complexity Classes V. More PCPs. Eric Rachlin

NP Complete Problems. COMP 215 Lecture 20

Some Algebra Problems (Algorithmic) CSE 417 Introduction to Algorithms Winter Some Problems. A Brief History of Ideas

NP Completeness. CS 374: Algorithms & Models of Computation, Spring Lecture 23. November 19, 2015

Complexity Theory of Polynomial-Time Problems

Automata Theory CS Complexity Theory I: Polynomial Time

CS151 Complexity Theory. Lecture 1 April 3, 2017

CSC 5170: Theory of Computational Complexity Lecture 4 The Chinese University of Hong Kong 1 February 2010

CS151 Complexity Theory

Towards Hardness of Approximation for Polynomial Time Problems

6.080 / Great Ideas in Theoretical Computer Science Spring 2008

Completeness for First-Order Properties on Sparse Structures with Algorithmic Applications

2 P vs. NP and Diagonalization

Lecture 5: The Landscape of Complexity Classes

Lecture 22: Counting

Essential facts about NP-completeness:

Complexity Theory VU , SS The Polynomial Hierarchy. Reinhard Pichler

Outline. Complexity Theory EXACT TSP. The Class DP. Definition. Problem EXACT TSP. Complexity of EXACT TSP. Proposition VU 181.

Complexity, P and NP

Some Algebra Problems (Algorithmic) CSE 417 Introduction to Algorithms Winter Some Problems. A Brief History of Ideas

Easy Problems vs. Hard Problems. CSE 421 Introduction to Algorithms Winter Is P a good definition of efficient? The class P

NP Completeness and Approximation Algorithms

1 Primals and Duals: Zero Sum Games

1 Randomized reduction: a first start

CS 583: Algorithms. NP Completeness Ch 34. Intractability

Notes on Computer Theory Last updated: November, Circuits

More Consequences of Falsifying SETH and the Orthogonal Vectors Conjecture [Full version]

NP-Completeness. Subhash Suri. May 15, 2018

8. INTRACTABILITY I. Lecture slides by Kevin Wayne Copyright 2005 Pearson-Addison Wesley. Last updated on 2/6/18 2:16 AM

Computability and Complexity Theory: An Introduction

Complexity Theory of Polynomial-Time Problems

Non-Deterministic Time

ITCS:CCT09 : Computational Complexity Theory Apr 8, Lecture 7

Theory of Computation Time Complexity

NP-Hardness reductions

CSC 5170: Theory of Computational Complexity Lecture 9 The Chinese University of Hong Kong 15 March 2010

COP 4531 Complexity & Analysis of Data Structures & Algorithms

NP and Computational Intractability

CISC 4090 Theory of Computation

Discrete Optimization 2010 Lecture 10 P, N P, and N PCompleteness

P, NP, NP-Complete, and NPhard

NP Completeness and Approximation Algorithms

-bit integers are all in ThC. Th The following problems are complete for PSPACE NPSPACE ATIME QSAT, GEOGRAPHY, SUCCINCT REACH.

Intro to Theory of Computation

CS256 Applied Theory of Computation

Computational Intractability 2010/4/15. Lecture 2

NP-complete problems. CSE 101: Design and Analysis of Algorithms Lecture 20

1 Reductions and Expressiveness

Design and Analysis of Algorithms

1 Computational problems

Lecture 17: Cook-Levin Theorem, NP-Complete Problems

Complexity (Pre Lecture)

Space and Nondeterminism

Friday Four Square! Today at 4:15PM, Outside Gates

Algorithms and Theory of Computation. Lecture 22: NP-Completeness (2)

1 PSPACE-Completeness

Polynomial-Time Reductions

UC Berkeley CS 170: Efficient Algorithms and Intractable Problems Handout 22 Lecturer: David Wagner April 24, Notes 22 for CS 170

7.8 Intractability. Overview. Properties of Algorithms. Exponential Growth. Q. What is an algorithm? A. Definition formalized using Turing machines.

CSE 105 THEORY OF COMPUTATION

Complexity Theory of Polynomial-Time Problems

Lecture 13, Fall 04/05

Lecture 22: PSPACE

NP-Complete problems

Show that the following problems are NP-complete

Reductions to Graph Isomorphism

Umans Complexity Theory Lectures

Lecture 18: More NP-Complete Problems

NP-Completeness Review

Transcription:

A FINE-GRAINED APPROACH TO ALGORITHMS AND COMPLEXITY Virginia Vassilevska Williams Stanford University

THE CENTRAL QUESTION OF ALGORITHMS RESEARCH ``How fast can we solve fundamental problems, in the worst case? etc.

ALGORITHMIC TECHNIQUES Divide and Conquer Kernelization Dynamic Programming Color-coding Greedy approaches Iterative compression Inclusionexclusion Backtracking Linear programming Semidefinite Programming Sum of Squares Specialized data structures

HARD PROBLEMS For many problems, the known techniques get stuck: Very important computational problems from diverse areas They have simple, often brute-force, classical algorithms No improvements in many decades

Example 1 A CANONICAL HARD PROBLEM k-sat Input: variables x 1,,x n and a formula F = C 1 C 2 C m so that each C i is of the form {y 1 y 2 y k } and i, y i is either x t or x t for some t. Output: A boolean assignment to {x 1,,x n } that satisfies all the clauses, or NO if the formula is not satisfiable Brute-force algorithm: try all 2 n assignments Best known algorithm: O(2 n-(cn/k) n d ) time for const c,d Goes to 2 n as k grows.

Example 2??? ANOTHER HARD PROBLEM: LONGEST COMMON SUBSEQUENCE (LCS) Given two strings on n letters ATCGGGTTCCTTAAGGG AT ATTGGTACCTTCAGGT GG_TACCTTCA_GG Find a subsequence of both strings of maximum length. Algorithms: Applications both in computational biology and in spellcheckers. Classical O(n 2 ) time Best algorithm: O(n 2 / log 2 n) time [MP 80] Solved daily on huge strings (Human genome: 3 x 10 9 base pairs.)

THE REAL WORLD AND NP-HARD PROBLEMS LIKE K-SAT I ve got data. I want to solve this algorithmic problem but I m stuck Ok, thanks, I feel better that none of my attempts worked. I ll use some heuristics. I m sorry, this problem is NP-hard. A fast algorithm for it would resolve a hard problem in CS/math.

THE REAL WORLD AND EASIER PROBLEMS LIKE LCS I ve got data. I want to solve this algorithmic problem but I m stuck But my data size n is huge Don t you have a faster algorithm??? Should I wait? Or should I be satisfied with heuristics? Great news Your problem is in P. Here s an O(n 2 ) time algorithm Uhm, I don t know This is already theoretically fast For some reason I can t come up with a faster algorithm for it right now

This is for a variety of reasons. IN THEORETICAL CS, POLYNOMIAL TIME = EFFICIENT/EASY. E.g. composing two efficient algorithms results in an efficient algorithm. Also, model-independence. However, noone would consider an O(n 100 ) time algorithm efficient in practice. If n is huge, then O(n 2 ) can also be inefficient.

No N 2 - ε time algorithms known for: WE ARE STUCK ON MANY PROBLEMS, EVEN JUST IN O(N 2 ) TIME Many string matching problems: Edit distance, Sequence local alignment, LCS, jumbled indexing General form: given two sequences of length n, how similar are they? All variants can be solved in O(n 2 ) time by dynamic programming. ATCGGGTTCCTTAAGGG ATTGGTACCTTCAGG

No N 2 - ε time algorithms known for: Many string matching problems WE ARE STUCK ON MANY PROBLEMS, EVEN JUST IN O(N 2 ) TIME Many problems in computational geometry: e.g Given n points in the plane, are any three colinear? A very important primitive

No N 2 - ε time algorithms known for: Many string matching problems WE ARE STUCK ON MANY PROBLEMS, EVEN JUST IN O(N 2 ) TIME Many problems in computational geometry Many graph problems in sparse graphs: e.g. Given an n node, O(n) edge graph, what is its diameter? Fundamental problem. Even approximation algorithms seem hard

No N 2 - ε time algorithms known for: Many string matching problems WE ARE STUCK ON MANY PROBLEMS, EVEN JUST IN O(N 2 ) TIME Many problems in computational geometry Many graph problems in sparse graphs Many other problems Why are we stuck? Are we stuck because of the same reason?

PLAN Traditional hardness in complexity A fine-grained approach New Developments

COMPUTATIONAL COMPLEXITY PSPACE NP P Logspace AC0 This traditional complexity class approach says little about runtime QBF with k alternations in 2 n-ω(k) time QBF gets easier as the # of quantifiers increases... Best SAT alg: 2 n Even O(n 2 ) time is inefficient Contains the 1000-clique problem; best runtime: Ω(n 790 )

N size of input NP P It also does not apply to problems in P Unless P=NP Theorem [Cook, Karp 72]: k-sat is NP-complete for all k 3. WHY IS K-SAT HARD? NP-completeness addresses runtime, but it is too coarsegrained That is, if there is an algorithm that solves k-sat instances on n variables in poly(n) time, then all problems in NP have poly(n) time solutions, and so P=NP. k-sat (and all other NP-complete problems) are considered hard because fast algorithms for them imply fast algorithms for many important problems.

TIME HIERARCHY THEOREMS For most natural computational models one can prove: for any constant c, there exist problems solvable in O(n c ) time but not in O(n c-ε ) time for any ε > 0. It is completely unclear how to show that a particular problem in O(n c ) time is not in O(n c-ε ) time for any ε > 0. Unconditional lower bounds seem hard. In fact, it is not even known if SAT is in linear time We instead develop a fine-grained theory of hardness that is conditional and mimics NP-completeness.

PLAN Traditional hardness in complexity A fine-grained approach Fine-grained reductions lead to new algorithms

Idea: Mimic NP-completeness FINE-GRAINED HARDNESS IDEA 1. Identify key hard problems 2. Reduce these to all (?) other problems believed hard 3. Hopefully form equivalence classes Goal: understand the landscape of algorithmic problems

CNF SAT IS CONJECTURED TO BE REALLY HARD Two popular conjectures about SAT on n variables [IPZ01]: ETH: 3-SAT requires 2 δ n time for some constant δ > 0. SETH: for every ε > 0, there is a k such that k-sat on n variables, m clauses cannot be solved in 2 (1- ε)n poly m time. So we can use k-sat as our hard problem and ETH or SETH as the conjecture we base hardness on.

Recent research [CGIMPS 16] suggests these problems are not equivalent Given a set S of n vectors in {0,1} d, for d = ω(log n) are there u, v 2 S with u v = 0? Conjecture: Orthog. Vecs. requires n 2-o(1) time. [W 04]: SETH implies this conjecture Easy O(n 2 d) time alg Best known [AWY 15]: n 2 - Θ(1 / log (d/log n)) Given a set S of n integers, are there a, b, c 2 S with a + b + c = 0? 3SUM Easy O(n 2 ) time alg [BDP 05]: ~n 2 / log 2 n time for integers [GP 14] : ~n 2 / log n time for reals Orthogonal vectors More key problems to blame Conjecture: 3SUM requires n 2-o(1) time. APSP Conjecture: APSP requires n 3-o(1) time. All pairs shortest paths: given an n-node weighted graph, find the distance between every two nodes. Classical algs: O(n 3 ) time [W 14]: n 3 / exp( log n) time

Author Runtime Year WORK ON APSP Floyd, Warshall n 3 1962 Fredman n 3 1976 Takaoka n 3 1992 Dobosiewicz n 3 1992 Classical problem Long history Han n 3 2004 Takaoka n 3 2004 Zwick n 3 2004 Chan n 3 2005 Han n 3 2006 Chan n 3 2007 Han, Takaoka n 3 2012 Williams n 3 2014

Idea: Mimic NP-completeness FINE-GRAINED HARDNESS 1. Identify key hard problems 2. Reduce these to all (?) other hard problems 3. Hopefully form equivalence classes Goal: understand the landscape of algorithmic problems

FINE-GRAINED REDUCTIONS A is (a,b)-reducible to B if for every ε>0 δ>0, and an O(a(n) 1-δ ) time algorithm that adaptively transforms any A-instance of size n to B-instances of size n 1,,n k so that Σ i b(n i ) 1-ε < a(n) 1-δ. If B is in O(b(n) 1-ε ) time, then A is in O(a(n) 1-δ ) time. Focus on exponents. We can build equivalences. Next: an example Intuition: a(n),b(n) are the naive runtimes for A and B. A reducible to B implies that beating the naive runtime for B implies also beating the naive runtime for A. B n A a(n) 1-δ B B B n 1, n 2,, n k

AN EXAMPLE FINE-GRAINED EQUIVALENCE THEOREM [VW 10]: Boolean matrix multiplication (BMM) is equivalent to Triangle detection under subcubic fine-grained reductions. BMM: Given two n x n Boolean matrices X and Y, return an n x n matrix Z where for all i and j, Z[i, j] = OR k (X[i, k] AND Y[k, j]). Triangle detection: Given an n node graph G, does it contain three vertices a, b, c, such that (a, b), (b, c), (c, a) are all edges? a We will show that (1) an O(n 3-e ) time alg for BMM can give an O(n 3-e ) time triangle alg, and (2) an O(n 3-e ) time alg for triangle can give an O(n 3-e/3 ) time BMM alg. b c

BMM: Given two n x n Boolean matrices X and Y, return an n x n matrix Z where for all i and j, Z[i, j] = OR k (X[i, k] AND Y[k, j]). If one can multiply Boolean matrices in O(n c ) time, then one can find a triangle in a graph in O(n c ) time. BMM CAN SOLVE TRIANGLE (ITAI, RODEH 1978) G=(V,E) - n node graph. A n x n adjacency matrix: for all pairs of nodes u,v A[u, v] = 1 if (u, v) is an edge and 0 otherwise. Say Z = Boolean product of A with itself. Then for all pairs of nodes u w, v Z[u, w] = OR v (A[u, v] AND A[v, w]) = 1 if there is a path of length 2 from u to w. and u w 0 otherwise. So G has a triangle iff there is some edge (u, w) in G s.t. Z[u, w] = 1.

BMM: Given two n x n Boolean matrices X and Y, return an n x n matrix Z where for all i and j, Z[i, j] = OR k (X[i, k] AND Y[k, j]). Reduction from BMM to triangle finding: t = n 2/3 A: rows of X, B: cols of X and rows of Y, C: cols of Y TRIANGLE CAN SOLVE BMM (VW 10) Split A into pieces A 1,,A t of size n/t Split B into pieces B 1,,B t of size n/t Split C into pieces C 1,,C t of size n/t Place an edge between every i in A and every j in C Z all zeros matrix For all triples A p,b q,c r in turn: While A p [ B q [ C r has a triangle, Let (i, j, k) be a triangle in A p [ B q [ C r Set Z[i, j] = 1 Remove (i, j) from the graph. A 1 A 2 A t-1 A t i a A ALL EDGES X[i, k]=1 X[a, b]=0 Graph representation of BMM B k b Y[k, j]=1 B 1 B 2 B t-1 B t j C C 1 C 2 C t-1 C t

Z all zeros matrix For all triples A p,b q,c r in turn: While A p [ B q [ C r has a triangle, Let (i, j, k) be a triangle in A p [ B q [ C r Set Z[i, j] = 1 Remove (i, j) from the graph. BMM TO TRIANGLE REDUCTION Correctness: Every triple of nodes i, j, k appears in some examined A p [ B q [ C r Runtime: Every call to the Triangle finding algorithm is due to either (1) Setting an entry Z[i, j] to 1, or (2) Determining that some triple A p [ B q [ C r doesn t have any more triangles this happens at most once per pair i, j this happens at most once per triple A p [ B q [ C r If the runtime for detecting a triangle is T(n) = O(n 3-ε ), then the reduction time is (n 2 + t 3 ) T(n/t). Setting t=n 2/3, we get: O(n 3 - ε/3 ).

Idea: Mimic NP-completeness FINE-GRAINED HARDNESS 1. Identify key hard problems 2. Reduce these to all (?) other hard problems 3. Hopefully form equivalence classes Goal: understand the landscape of algorithmic problems

Using other hardness assumptions, one can unravel even more structure N input size n number of variables or vertices SOME STRUCTURE WITHIN P Sparse graph diameter [RV 13], approximate graph eccentricities [AVW 16], local alignment, longest common substring* [AVW 14], Frechet distance [Br 14], Edit distance [BI 15], LCS, Dynamic time warping [ABV 15, BrK 15], subtree isomorphism [ABHVZ 15], N 2- ε Many dynamic problems [P 10],[AV 14], [HKNS 15], [RZ 04] N 2- ε Huge literature in comp. geom. [GO 95, BHP98, ]: Geombase, 3PointsLine, 3LinesPoint, Polygonal Containment String problems: Sequence local alignment [AVW 14], jumbled indexing [ACLL 14] N 2- ε 2 (1 - δ)n 3SUM N 2- ε Orthog. vectors k-sat 8 k [W 04] APSP n 3- ε N 1.5- ε equivalent N 1.5- ε n 3- ε In dense graphs: radius, median, betweenness centrality [AGV 15], negative triangle, second shortest path, replacement paths, shortest cycle [VW 10],

PLAN Traditional hardness in complexity A fine-grained approach New developments The quest for more believable conjectures

THE QUEST FOR MORE PLAUSIBLE CONJECTURES Two problems harder than CNF-SAT,3SUM, and APSP Longest common subsequence, Formula SAT and Branching Programs

[Abboud-VW-Yu STOC 15]: Two hard problems for node-colored graphs (2 n ) 1- ε n 2- ε 3SUM k-sat for all k APSP n 3- ε D n 3- ε Matching Triangles Given an n-node graph G, a color for every vertex in G, and an integer D, is there a triple of colors q1,q2,q3 such that there are at least D triangles in G with node colors exactly q1,q2,q3? Triangle Collection n 3- ε Given an n-node graph G and a color for every vertex in G, is there a triple of colors q1,q2,q3 such that there are no triangles in G with node colors exactly q1,q2,q3? 1. Graphs don t have weights, just node colors 2. Any NO reduction from these problems would imply hardness under all three conjectures

SOME STRUCTURE WITHIN Dynamic P Sparse graph diameter [RV 13], local alignment, longest common substring* [AVW 14], Frechet distance [Br 14], Edit distance [BI 15], LCS, Dynamic Time Warping [ABV 15, BrK 15] N 2- ε problems [P 10],[AV 14], [HKNS 15], [RZ 04] N 1.5- ε N 2- ε Huge literature in comp. geom. [GO 95, BH-P98, ]: Geombase, 3PointsLine, 3LinesPoint, Polygonal Containment String problems: Sequence local alignment [AVW 14], jumbled indexing [ACLL 14] N 2- ε 2 (1 - δ)n 3SUM N 2- ε Orthog. vectors k-sat 8 k N 1.5- ε APSP n 3- ε equivalent n 3- ε In dense graphs: radius, median, betweenness centrality [AGV 15], negative triangle, second shortest path, replacement paths, shortest cycle [VW 10], Triangle collection* n 3- ε S-T max flow, dynamic max flow, [AVY 15]

THE QUEST FOR MORE PLAUSIBLE CONJECTURES Two problems harder than CNF-SAT,3SUM, and APSP Longest common subsequence, Formula SAT and Branching Programs

CIRCUIT-STRONG-ETH The most successful hypothesis has been SETH It is ultimately about SAT of linear size CNF-formulas There are more difficult satisfiability problems: CIRCUIT-SAT C-SETH: satisfiability of circuits from NC-SAT circuit class C on n variables and NC1-SAT size s requires 2 n-o(n) poly(s) time. [Williams 10, 11]: a 2 n /n 10 time SAT algorithm implies circuit lower bounds for C (for E NP and others); the bigger the class the stronger the lower bound

A VERY RECENT DEVELOPMENT Theorem [Abboud-Hansen-VW-Williams 16]: There is an efficient reduction from Satisfiability for non-deterministic branching programs (BPs) of size T and width W and n input variables to the following string problems on strings of length N = 2 n/2 T O(log W) : Longest Common Subsequence, Edit Distance, Dynamic Time Warping, etc.

THEOREM [ABBOUD-HANSEN-VW-WILLIAMS 15]: THERE IS AN EFFICIENT REDUCTION FROM SATISFIABILITY FOR NON-DETERMINISTIC BRANCHING PROGRAMS (BPS) OF SIZE T AND WIDTH W AND N INPUT VARIABLES TO THE FOLLOWING STRING PROBLEMS ON STRINGS OF LENGTH N = 2 N/2 T O(LOG W) : LONGEST COMMON SUBSEQUENCE, EDIT DISTANCE, DYNAMIC TIME WARPING, ETC. A type of reachability question. Proof encodes a Savitch-like construction into the LCS/Edit distance instance. W x 1 x 3 x 2 x 1 0 0 1 s 1 1 1 0 0 0 0 1 0 1 1 1 0 0 1 T t BP: edge-labelled, directed, layered graph. Start node s, accept node t. Width: W nodes per layer. Size: T layers. Each layer labeled with a variable. A variable can label many layers. Each edge labeled with 0 or 1. An input 001 is accepted if it generates an s-t path.

A VERY RECENT DEVELOPMENT Theorem [Abboud-Hansen-VW-Williams 16]: There is an efficient reduction from Satisfiability for non-deterministic branching programs (BPs) of size T and width W and n input variables to the following string problems on strings of length N = 2 n/2 T O(log W) : Longest Common Subsequence, Edit Distance, Dynamic Time Warping, etc. [Barrington 85]: BPs with T=2 polylog n and W=5 capture NC. The above problems require N 2-o(1) time under NC-SETH.

MORE CONSEQUENCES OF BP-SAT FOR N,T,W EDIT DISTANCE ETC. ON 2 N/2 T O(LOG W) Impressive SAT algs non- BPs with T=2 o(sqrt n) and W=2 o(sqrt n) can represent any deterministic Turing machine using o(sqrt n) space Edit Distance (or LCS etc) in O(n 2-ε ) time implies a nontrivial improvement over exhaustive search for checking SAT of complex objects that can easily implement e.g. cryptographic primitives Much more surprising than refuting SETH

MORE CONSEQUENCES OF BP-SAT FOR N,T,W EDIT DISTANCE ETC. ON 2 N/2 T O(LOG W) Circuit Lower Bounds If Edit Distance (or LCS etc) has O(n 2-ε ) time algorithms for any ε > 0, then E NP does not have: Non-uniform 2 o(n) -size Boolean Formulas we don t even know if the enormous Σ 2 EXP has 2 o(n) -size depth-3 circuits Non-uniform o(n)-depth circuits of bouded fan-in Non-uniform 2 o(sqrt(n)) -size non-deterministic branching programs

A POLYLOG SHAVED IS A LOWER BOUND MADE n 2 /log 2.1 n? [Williams 14, Abboud-Williams-Yu 15]: APSP can be solved in n 3 / log ω(1) n time, OV can be solved in n 2 /log ω(1) n time Does Edit Distance (or LCS etc) have such an algorithm? (The current best algorithms run in ~ n 2 /log 2 n time.) PARTIAL ANSWER: An n 2 / log ω(1) n algorithm for Edit Distance (or LCS etc) implies that E NP is not in NC1. Also meaningful for particular polylogs. E.g. if Edit Distance (or LCS etc) has an n 2 /log 100 n time algorithm, then E NP does not have non-uniform Boolean formulas of size n 5.

Thank you