# HG changeset patch
# User Christian Urban
Description:
Regular expressions
- are extremely useful for many text-processing tasks such as finding patterns in texts,
+ are extremely useful for many text-processing tasks, such as finding patterns in texts,
lexing programs, syntax highlighting and so on. Given that regular expressions were
introduced in 1950 by Stephen Kleene,
you might think regular expressions have since been studied and implemented to death. But you would definitely be
mistaken: in fact they are still an active research area. For example
this paper
about regular expression matching and derivatives was presented just last summer at the international
- FLOPS'14 conference. The task in this project is to implement their results.
The background for this project is that some regular expressions are “evil” @@ -106,9 +106,10 @@ (for example subexpression matching, which my rainy-afternoon matcher cannot). I am sure they thought about the problem much longer than a single afternoon. The task in this project is to find out how good they actually are by implementing the results from their paper. - Their approach is based on the concept of derivatives. - I used them once myself in a paper - in order to prove the Myhill-Nerode theorem. + Their approach to regular expression matching is also based on the concept of derivatives. + I used derivatives very successfully once for something completely different in a + paper + about the Myhill-Nerode theorem. So I know they are worth their money. Still, it would be interesting to actually compare their results with my simple rainy-afternoon matcher and potentially “blow away” the regular expression matchers in Python and Ruby (and possibly in Scala too). The application would be to implement a fast lexer for @@ -623,7 +624,7 @@
- Last modified: Sun Nov 9 21:37:30 GMT 2014 + Last modified: Sun Nov 9 21:47:12 GMT 2014 [Validate this page.]