equal
deleted
inserted
replaced
10 { println(s" Problem with: $url"); ""} |
10 { println(s" Problem with: $url"); ""} |
11 } |
11 } |
12 |
12 |
13 // regexes for URLs, for "my" domain and for email addresses |
13 // regexes for URLs, for "my" domain and for email addresses |
14 val http_pattern = """"https?://[^"]*"""".r |
14 val http_pattern = """"https?://[^"]*"""".r |
15 val email_pattern = """([a-z0-9_\.-]+)@([\da-z\.-]+)\.([a-z\.]{2,6})""".r |
15 val email_pattern = """([a-z0-9_\.-]+)@([\da-z\.-]+)\.([a-z\.]{2,6})""".r (*@\label{emailline}@*) |
16 |
16 |
17 def unquote(s: String) = s.drop(1).dropRight(1) |
17 def unquote(s: String) = s.drop(1).dropRight(1) |
18 |
18 |
19 def get_all_URLs(page: String) : Set[String] = |
19 def get_all_URLs(page: String) : Set[String] = |
20 http_pattern.findAllIn(page).map(unquote).toSet |
20 http_pattern.findAllIn(page).map(unquote).toSet |
25 def crawl(url: String, n: Int) : Unit = { |
25 def crawl(url: String, n: Int) : Unit = { |
26 if (n == 0) () |
26 if (n == 0) () |
27 else { |
27 else { |
28 println(s"Visiting: $n $url") |
28 println(s"Visiting: $n $url") |
29 val page = get_page(url) |
29 val page = get_page(url) |
30 print_str(email_pattern.findAllIn(page).mkString("\n")) |
30 print_str(email_pattern.findAllIn(page).mkString("\n")) (*@\label{mainline}@*) |
31 for (u <- get_all_URLs(page).par) crawl(u, n - 1) |
31 for (u <- get_all_URLs(page).par) crawl(u, n - 1) |
32 } |
32 } |
33 } |
33 } |
34 |
34 |
35 // staring URL for the crawler |
35 // staring URL for the crawler |