equal
  deleted
  inserted
  replaced
  
    
    
   247   case ("pl", _) => T_LPAREN | 
   247   case ("pl", _) => T_LPAREN | 
   248   case ("pr", _) => T_RPAREN | 
   248   case ("pr", _) => T_RPAREN | 
   249 }  | 
   249 }  | 
   250   | 
   250   | 
   251   | 
   251   | 
   252 def tokenise(s: String) : List[Token] =   | 
   252 def tokenise(s: String) : List[Token] = { | 
   253   lexing_simp(FUN_REGS, s).collect(token)  | 
   253   val tks = lexing_simp(FUN_REGS, s).collect(token)  | 
         | 
   254   if (tks.length != 0) tks  | 
         | 
   255   else { println (s"Tokenise Error") ; sys.exit(-1) }      | 
         | 
   256 }  | 
   254   | 
   257   | 
   255 def serialise[T](fname: String, data: T) = { | 
   258 def serialise[T](fname: String, data: T) = { | 
   256   import scala.util.Using  | 
   259   import scala.util.Using  | 
   257   Using(new ObjectOutputStream(new FileOutputStream(fname))) { | 
   260   Using(new ObjectOutputStream(new FileOutputStream(fname))) { | 
   258     out => out.writeObject(data)  | 
   261     out => out.writeObject(data)  |