equal
deleted
inserted
replaced
269 val tks = lexing_simp(FUN_REGS, s).collect(token) |
269 val tks = lexing_simp(FUN_REGS, s).collect(token) |
270 if (tks.length != 0) tks |
270 if (tks.length != 0) tks |
271 else { println (s"Tokenise Error") ; sys.exit(-1) } |
271 else { println (s"Tokenise Error") ; sys.exit(-1) } |
272 } |
272 } |
273 |
273 |
274 import ammonite.ops._ |
274 //import ammonite.ops._ |
275 |
275 |
276 //@doc("Tokenising a file.") |
276 //@doc("Tokenising a file.") |
277 @main |
277 @main |
278 def main(fname: String) = { |
278 def main(fname: String) = { |
279 println(tokenise(os.read(os.pwd / fname))) |
279 println(tokenise(os.read(os.pwd / fname))) |