in reply to Re^3: Block-structured language parsing using a Perl module?
in thread Block-structured language parsing using a Perl module?
you have to know enough theory to know both what type of parser you can use on a grammar and if your grammar is even parsable.
Hm. That would be a justification for it, but so far, none of the modules discussion even begins to allow you to answer those types of questions.
I can't speak about the performance of Regexp::Grammars, but if I were doing something like this, I'd start there for ease of use.
Hm. I'm going through the docs for Regexp::Grammars now, and trying to do so with an open mind, but honestly, what I'm reading is making my skin crawl.
The questions I am asking myself at this point are:
- Why do I need to add/remove directives to my grammar in order to enable/disable debugging?
For the same reason that these errors exist. Because the module trades a conventional interface for overloading qr//.
The POD devotes a section to restricting the scope of the module's effect, by wrapping it use in a do block to prevent it from messing with other qr//s in the program. A can see no advantages to the cutesy interface over use Regexp::Grammars qw[ compile ]; my $re = compile $grammar;; and at least 3 disadvantages.
- Why complicate the interface by adding a logfile option?
Just send errors/debug/and trace to STDERR and let me redirect -- manually or programmically wherever I want it.
- Dumping 3 gigatonnes of trace to a file does not constitute a debugger and is no substitute for decent diagnostics.
- Why are 3 aliases for everything; and 3 ways to do everything including creating aliases?
- Why do I need both a subroutine-based interface and an OO interface?
What benefit is there to an OO interface?
Does it avoid the use of globals like %MATCH and %/ etc? No.
Can I run two instances of a grammar concurrently? No. The regex engine wouldn't allow it.
It's just pseud-OO.
I'd use Marpa for speed and completeness.
The trouble with Marpa is that it only does half the job. You have to tokenise the source text yourself, and then feed it to the parser in labeled chunks.
By the time you've written the code to tokenise the input, and then recognise the tokens so you can label them for the "parser", one wonders what parsing there is left for the "parser" to do.
Its like buying a dog and barking yourself.