Notes on parsing theory, part 3

stereobooster - Jan 29 '21 - - Dev Community

Based on Theory of Computation (CS3102), Spring 2017, videos are here.

Definitions

Let’s start with refinement of terminology from part 1.

Alphabet (Σ - sigma) is a finite, non-empty set of symbols letters. Example: Σ = {a, b}.

String (w) is a finite sequence of symbols letters chosen from Σ. Example: ababba.

Language (L) a (possibly infinite) set of strings. Example: L = {a, aa, aaa}.

String length is a number of letters in it: |aba| = 3.

String concatenation : w₁w₂. Example: ab•ba = abba.

Empty string : ε (epsilon). |ε| = 0, ∀w w•e = e•w = w ( - for all).

Language concatenation : L₁L₂ = {w₁w₂ | w₁∈L₁, w₂∈L₂} - Cartesian product of two sets ( - in). Example: {1,2}•{a,b,...} = {1a,2a,1b,2b,...}

String exponentiation : wᵏ = ww...w (k times). Example: a³=aaa

Language exponentiation : Lᵏ = LL...L (k times). LL = L², Lᵏ = LLᵏ⁻ⁱ, L⁰ = {e}. Example: {0,1}³² - all binary words of length of 32.

String reversal : wᴿ. Example: (aabc)ᴿ=cbaa.

Language reversal : Lᴿ = {wᴿ | w∈L}. Example: {ab, cd}ᴿ = {ba, dc}.

Language union : L₁⋃L₂ = {w | w∈L₁ or w∈L₂} - set union. Example: {a}⋃{b,aa} = {a,b,aa}.

Language intersection : L₁∩L₂ = {w | w∈L₁ and w∈L₂} - set intersection. Example: {a,b}∩{b,c} = {b}.

Language difference : L₁ - L₂ = {w | w∈L₁ and w∉L₂} - set difference ( not in). Example: {a,b} - {b,d} = {a}.

Kleene closure : L* = L⁰⋃Lⁱ⋃L²…, L+ = Lⁱ⋃L²⋃L³…. Example {a}* = {ε,a,aa,aaa…}, {a}+ = {a,aa,aaa…}

All finite strings (over Σ): Σ*, L⊆Σ* ∀L. Note : Robins doesn’t provide a rule for alphabet exponentiation, so here is how I interpret this: If is a language which contains all strings of length 1 over alpahbet Σ (ℒ=Σ) then L⊆ℒ* ∀L.

“Trivial” language : {ε}. {ε}•L = L•{ε} = L.

Empty language : Ø. Ø* = {ε}.

Theorem: Σ* contains only finite string (but posibliy infinite number of strings).

Theorem: (L*)* = L*

Theorem: L+ = LL*.

The extended Chomsky Hierarchy

Originally Chomsky proposed 4-types of grammar classes. Since then hierarchy was refined with the more granular division.

I can now understand some notations on this diagram aⁿbⁿcⁿ (a string which contains equal size sequences of a, b, c), wwᴿ (palindrome), aⁿbⁿ, a* (Kleene star), {a, b} finite set.

But what about P, NP, LBA, etc? Those notations come from 3 different branches:

  • Computability theory: What problems are solvable for computers?
    • Decidable vs undecidable (Turing degrees)
  • Complexity theory: What makes some problems easy or hard for computers?
    • P, NP, PSPACE, EXPTIME, EXPSPACE
  • Automata theory: What formal model of computation is required for a problem?
    • LBA (linear bounded automaton), DPA (deterministic pushdown automaton) etc.

Automata theory

Notes based on Automata Theory course from Stanford.

Automatons are abstract models of machines that perform computations on an input by moving through a series of states or configurations. At each state of the computation, a transition function determines the next configuration on the basis of a finite portion of the present configuration. As a result, once the computation reaches an accepting configuration, it accepts that input. The most general and powerful automata is the Turing machine.

The behavior of these discrete systems is determined by the way that the system is constructed from storage and combinational elements. Characteristics of such machines include:

  • Inputs : assumed to be sequences of symbols letters selected from a finite set of input signals (string w∈Σ*).
  • Outputs : sequences of symbols letters selected from a finite set Z.
  • States : finite set Q, whose definition depends on the type of automaton.

There are four major families of automaton :

  • Finite-state machine (FSM) or finite automata (automaton)
    • deterministic finite automata (DFA) and non-deterministic FA (NFA)
  • Pushdown automata (PDA)
    • Deterministic PDA or non-deterministic PDA (NPDA, NPA)
  • Linear-bounded automata (LBA)
  • Turing machine (TM)

Finite-state machine

FSM is a “machine” that changes states while processing symbols, one at a time. FSM is a 5-tuple M=(Q, Σ, δ, q₀, F) (M for a machine)

  • Finite set of states: Q = {q₀, q₁, q₂, ..., qₓ}
  • Transition function: δ: Q×Σ → Q
  • Initial state: q₀∈Q
  • Final states: F⊆Q (is a subset of Q)

Acceptance : end up in a final state (F).

Rejection : anything else (including hang-up/crash)

For example: automata that accepts abababab…= (ab)*.

M=({q₀,q₁}, {a,b}, {((q₀,a),q₁), ((q₁,b),q₀)}, q₀, {q₀})

Enter fullscreen mode Exit fullscreen mode

As a graph:

× is a Cartesian product, is a mapping (function or more generally relationship). As a result of the Cartesian product of two sets, we get a set of tuples. We can think of tuples as 2-d coordinates, so we can plot the function as 3-d scatter plot or as a table:

a b
→q₀ q₁
q₁ q₀

For example, state changes for accepting string abab:

But sometimes it is drawn like this (this is confusing IMO):

This machine will “crash” on input string “abba”. To fix this we can add more edges (make machine total):

We can have unconnected nodes (q₃ is final state, but it’s unreachable from start state q₀):

If we extend transition function to strings (from symbols) δ: Q×Σ* → Q we can represent machine compactly.

δ(q₀,ww₁) = δ(δ(q₀,w),w₁), where δ(qₓ,ε) = qₓ.

Example 1: δ(q₀,ab) = δ(δ(q₀,a),b)

Example 2:

Note: {a,b} = PCRE /^[ab]$/

Language of M is L(M) = {w∈Σ* | δ(q₀,w)∈F}

Language is regular iff (if and only if) it is accepted by some FSM.

Note : this is one more way to formally specify language, the first one was with grammar (L(G))

Non-detemenistics FSM

Non-determinism: generalizes determinism, where many “next moves” are allowed at each step. New definition of transition function:

δ: 2Q×Σ → 2Q
Enter fullscreen mode Exit fullscreen mode

Computation becomes a “tree”. Acceptance : any path from the root (start state) to some leaf (a final state).

Example

A machine that accepts regular language in which each word has b as the second symbol before the end. L = {ba, bb, aba, abb, bba, bbb ...}.

As regular expression: {a,b}*b{a,b}. AS PCRE (Perl Compatible Regular Expressions): /^[ab]*b[ab]$/. As a graph:

As a table:

a b
→q₀ q₀ q₀, q₁
q₁ q₂ q₂
q₂

Note that we have two possible states in the top right cell. This is a nondeterministic choice - with the same input machine can be in two different states.

The machine reads symbols of the string from left to right.

Let’s see state transitions for string “bb”.

Let’s see state transitions for string “bbb”.

Theorem : Non-determinism in FAs doesn’t increase power.

Proof by simulation:

  • Construct all super-states, one per each state subset.
  • New super-transition function jumps among super-states, simulating the old transition function
  • Initial super state is those containing the old initial state.
  • Final super states are those containing old final states.
  • Resulting DFA accepts the same language as the original NFA but can have exponentially more states.

Let’s convert previously defined NFA to DFA. We take the initial transition table and convert the first row to sets:

a b
→{q₀} {q₀} {q₀, q₁}

Now we have one new state {q₀, q₁} let’s construnct transitions for it:

a b
{q₀, q₁} {q₀,q₂} {q₀, q₁, q₂}

Continue:

a b
{q₀, q₂} {q₀} {q₀, q₁}
{q₀, q₁, q₂} {q₀, q₂} {q₀, q₁, q₂}

Let’s draw a graph:

Let’s see state transitions for string “bb”:

Let’s see state transitions for string “bbb”:

Read more about conversion of NFA to DFA here.

As we can see NFAs and DFAs have the same computational power, but NFAs are more compact and “clearly communicate intention”.

Regular Expressions

Regular expressions are used to denote regular languages. They can represent regular languages and operations on them compactly. The set of regular expressions over an alphabet Σ is defined recursively as below. Any element of that set is a regular expression.

1Regexp = Ø (* empty language *)
2 | {ε} (* trivial language *)
3 | {x} (* singleton language *)
4 | Regexp + Regexp (* union *)
5 | Regexp Regexp (* concatenation *)
6 | Regexp* (* Kleene closure *)

Enter fullscreen mode Exit fullscreen mode

Union R + S:

Note : Language union {a} ∪ {b} = {a,b} ~ RE union a + b ~ PCRE /^[ab]$/ or /^(a|b)$/.

Concatenation RS:

Kleene closure R*:

Theorem : Any regular expression is accepted by some FA.

A FA for a regular expression can be built by composition.

Note here suppose to be an example but I’m lazy to draw it. See original lecture slide on page 28.

FA Minimization

Merging nodes

Merging edges

Theorem [Hopcroft 1971]: the number N of states in a FA can be minimized within time O(N log N). Based on earlier work [Huffman 1954] & [Moore 1956].

Conjecture : Minimizing the number of states in a nondeterministic FA can not be done in polynomial time. (Is this because non-deterministic FA corresponds to deterministic FA with an exponential number of nodes?)

Theorem : Minimizing the number of states in a pushdown automaton (or TM) is undecidable.

Theorem : Any FA accepts a language denoted by some RE. Proof : use “generalized finite automata” where a transition can be a regular expression (not just a symbol), and: Only 1 super start state and 1 (separate) super final state. Each state has transitions to all other states (including itself), except the super start state, with no incoming transitions, and the super final state, which has no outgoing transitions.

Now reduce the size of the GFA by one stateat each step. A transformation step is as follows:

Such a transformation step is always possible, until the GFA has only two states, the super-start, and super-final states:

The label of the last remaining transition is the regular expression corresponding to the language of the original FA

Corollary : FAs and REs denote the same class of languages.

To be continued

Lectures from Robins are quite good…

I didn’t cover:

  • ε-transitions in NFAs
  • transformation and minimisation of FSMs
  • Regular Expressions Identities
  • Decidable Finite Automata Problems

See also:

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .