What complexity implications are there from not being in an abstract family of languages?
A classic result is that a if a language is not regular its decision problem requires more than constant memory (i.e. $\Omega(\log \log n)$).
I am wondering if there are similar results for other classes of language (e.g. context-free or tree-adjoining). Specifically results of the form
If a language is not a $X$, then it must have space/time complexity $\omega(f)$ (or $\Omega(f)$)
where $X$ is some abstract family of languages. Naturally I don't care about trivial results like "If a language is not $\mathrm{DSPACE}\left(n^2\right)$ then it requires $\omega(n^2)$ space". I am particularly interested when $X$ is between regular and context-sensitive.
I had a look for myself, and I was of course able to find tons of results that being a particular class of formal language implies the existence of an algorithm of a certain complexity, but none more of the form like the above.
1 answer
Transform your statement
\[\text{If a language is not an $X$, then it must have space/time complexity $\omega(f)$ (or $\Omega(f)$)}\] into its contrapositive: \[\text{If a language has complexity $O(g), g < f$, then it is an X}\]Now perhaps it is easier to see why this problem tends to be very hard: every computational problem can be treated as a sort of "language". For instance $s$-$t$ connectivity can be thought of as a language in which graphs with two connected points are valid expressions and graphs with two disconnected points are invalid expressions. And this example suffices to show that even languages with space complexity $O(\log n)$ can be quite nontrivial.
Indeed, the only natural way (that I see) of solving such a problem is by making some universal "translation algorithm" from $O(g)$ languages to $X$ languages. I think the regex example you gave illustrates the difficulty with this approach well: it's easy to show that a regex needs only constant space complexity, but as soon as you add almost anything (say, backreferences) regexes become either pushdown automata or Turing-complete. (Of course, pushdown automata all have linear complexity, but not conversely---something with linear complexity would correspond to a "double-pushdown automaton" but I'm not sure if there's a conventional name for languages parsable by such automata!)
In short, I think it is precisely because of the versatility of, say, $O(n \log n)$ deterministic algorithms or $O(\log n)$ nondeterministic algorithms that it is very hard to correlate them precisely with natural concepts in formal grammar such as context-free or deterministic context-free languages, even though as you noted it tends to be possible to upper-bound space/time complexity given grammatical complexity.

0 comment threads