equal
  deleted
  inserted
  replaced
  
    
    
     1 // A simple lexer inspired by work of Sulzmann & Lu  | 
     1 // Detokenising the ouput of Tokeniser  | 
     2 //==================================================  | 
     2 //=====================================  | 
         | 
     3 //  | 
         | 
     4 // call with   | 
         | 
     5 //  | 
         | 
     6 //     scala detokenise.scala fib.tks  | 
         | 
     7 //  | 
         | 
     8 //     scala detokenise.scala loops.tks  | 
     3   | 
     9   | 
     4   | 
    10 object Detokenise { | 
     5 object Delexer { | 
         | 
     6   | 
    11   | 
     7 import java.io._  | 
    12 import java.io._  | 
     8   | 
    13   | 
     9 abstract class Token extends Serializable   | 
    14 abstract class Token extends Serializable   | 
    10 case object T_SEMI extends Token  | 
    15 case object T_SEMI extends Token  | 
    22   in.close  | 
    27   in.close  | 
    23   data  | 
    28   data  | 
    24 }  | 
    29 }  | 
    25   | 
    30   | 
    26 def main(args: Array[String]) = { | 
    31 def main(args: Array[String]) = { | 
    27   println("TEST\n" ++ deserialise[List[Token]]("/tmp/nflx").mkString)   | 
    32   val fname = args(0)  | 
         | 
    33   val tks = deserialise[List[Token]](fname)  | 
         | 
    34   println(s"Reading back from ${fname}:\n${tks.mkString("\n")}")   | 
    28 }  | 
    35 }  | 
    29   | 
    36   | 
    30   | 
    37   | 
    31 }  | 
    38 }  |