progs/fun_tokens.scala
changeset 655 689d1bdd91c0
parent 645 8da29f6ef225
equal deleted inserted replaced
654:718a9913db60 655:689d1bdd91c0
   247   case ("pl", _) => T_LPAREN
   247   case ("pl", _) => T_LPAREN
   248   case ("pr", _) => T_RPAREN
   248   case ("pr", _) => T_RPAREN
   249 }
   249 }
   250 
   250 
   251 
   251 
   252 def tokenise(s: String) : List[Token] = 
   252 def tokenise(s: String) : List[Token] = {
   253   lexing_simp(FUN_REGS, s).collect(token)
   253   val tks = lexing_simp(FUN_REGS, s).collect(token)
       
   254   if (tks.length != 0) tks
       
   255   else { println (s"Tokenise Error") ; sys.exit(-1) }     
       
   256 }
   254 
   257 
   255 def serialise[T](fname: String, data: T) = {
   258 def serialise[T](fname: String, data: T) = {
   256   import scala.util.Using
   259   import scala.util.Using
   257   Using(new ObjectOutputStream(new FileOutputStream(fname))) {
   260   Using(new ObjectOutputStream(new FileOutputStream(fname))) {
   258     out => out.writeObject(data)
   261     out => out.writeObject(data)