blob: d47fe60492b3b4abbc254d61fb15ffc524c913f3 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
|
package sandcrawler
import com.twitter.scalding._
import parallelai.spyglass.base.JobBase
import parallelai.spyglass.hbase.{HBaseSource, HBasePipeConversions, HBaseConstants}
import parallelai.spyglass.hbase.HBaseConstants.SourceMode
import cascading.tuple.Fields
import cascading.property.AppProps
import java.util.Properties
class HBaseRowCountJob(args: Args) extends JobBase(args) with HBasePipeConversions {
// For now doesn't actually count, just dumps a "word count"
val output = args("output")
val hbs = new HBaseSource(
//"table_name",
//"quorum_name:2181",
"wbgrp-journal-extract-0-qa", // HBase Table Name
"mtrcs-zk1.us.archive.org:2181", // HBase Zookeeper server (to get runtime config info; can be array?)
new Fields("key"),
List("file"),
List(new Fields("size", "mimetype")),
sourceMode = SourceMode.SCAN_ALL)
.read
.debug
.groupAll { _.size('count) }
.write(Tsv(output))
}
|