aboutsummaryrefslogtreecommitdiffstats
path: root/scalding/src
diff options
context:
space:
mode:
authorEllen Spertus <ellen.spertus@gmail.com>2018-06-07 12:52:08 -0700
committerEllen Spertus <ellen.spertus@gmail.com>2018-06-07 12:52:08 -0700
commit6eca6290aa3fc829f4767023ae075350a0a78192 (patch)
tree9684b67ee0b2ac8e7ca3a047fc8368477faf19f5 /scalding/src
parentf81e193b4a2c7f61a46d1ead1bdf6b746997b3a8 (diff)
downloadsandcrawler-6eca6290aa3fc829f4767023ae075350a0a78192.tar.gz
sandcrawler-6eca6290aa3fc829f4767023ae075350a0a78192.zip
Added status count.
Diffstat (limited to 'scalding/src')
-rw-r--r--scalding/src/main/scala/sandcrawler/HBaseStatusCountJob.scala28
-rw-r--r--scalding/src/test/scala/sandcrawler/HBaseStatusCountTest.scala61
2 files changed, 89 insertions, 0 deletions
diff --git a/scalding/src/main/scala/sandcrawler/HBaseStatusCountJob.scala b/scalding/src/main/scala/sandcrawler/HBaseStatusCountJob.scala
new file mode 100644
index 0000000..0675efc
--- /dev/null
+++ b/scalding/src/main/scala/sandcrawler/HBaseStatusCountJob.scala
@@ -0,0 +1,28 @@
+package sandcrawler
+
+import cascading.property.AppProps
+import cascading.tuple.Fields
+import com.twitter.scalding._
+import java.util.Properties
+import parallelai.spyglass.base.JobBase
+import parallelai.spyglass.hbase.{HBaseSource, HBasePipeConversions}
+import parallelai.spyglass.hbase.HBaseConstants.SourceMode
+
+class HBaseStatusCountJob(args: Args) extends JobBase(args) with HBasePipeConversions {
+ val output = args("output")
+
+ HBaseStatusCountJob.getHBaseSource
+ .read
+ .fromBytesWritable(List('status))
+// .debug
+ .groupBy('status){group => group.size('count)}
+ .write(Tsv(output))
+}
+
+object HBaseStatusCountJob {
+ def getHBaseSource = HBaseBuilder.build(
+ "wbgrp-journal-extract-0-qa", // HBase Table Name
+ "mtrcs-zk1.us.archive.org:2181", // HBase Zookeeper server (to get runtime config info; can be array?)
+ List("grobid0:status"),
+ SourceMode.SCAN_ALL)
+}
diff --git a/scalding/src/test/scala/sandcrawler/HBaseStatusCountTest.scala b/scalding/src/test/scala/sandcrawler/HBaseStatusCountTest.scala
new file mode 100644
index 0000000..8b5c3d6
--- /dev/null
+++ b/scalding/src/test/scala/sandcrawler/HBaseStatusCountTest.scala
@@ -0,0 +1,61 @@
+package sandcrawler
+
+import cascading.tuple.{Tuple, Fields}
+import com.twitter.scalding.{JobTest, Tsv, TupleConversions}
+import org.apache.hadoop.hbase.io.ImmutableBytesWritable
+import org.apache.hadoop.hbase.util.Bytes
+import org.junit.runner.RunWith
+import org.scalatest.FunSpec
+import org.scalatest.junit.JUnitRunner
+import org.slf4j.LoggerFactory
+import parallelai.spyglass.hbase.HBaseSource
+import parallelai.spyglass.hbase.HBaseConstants.SourceMode
+import scala._
+
+@RunWith(classOf[JUnitRunner])
+class HBaseStatusCountTest extends FunSpec with TupleConversions {
+
+ val output = "/tmp/testOutput"
+
+ val log = LoggerFactory.getLogger(this.getClass.getName)
+
+ val statusType1 = "200"
+ val statusType2 = "404"
+
+ val sampleData = List(
+ List("sha1:K2DKSSVTXWPRMFDTWSTCQW3RVWRIOV3Q", statusType1),
+ List("sha1:C3YNNEGH5WAG5ZAAXWAEBNXJWT6CZ3WU", statusType1),
+ List("sha1:SDKUVHC3YNNEGH5WAG5ZAAXWAEBNX4WT", statusType2),
+ List("sha1:35985C3YNNEGH5WAG5ZAAXWAEBNXJW56", statusType2),
+ List("sha1:885C3YNNEGH5WAG5ZAAXWA8BNXJWT6CZ", statusType2),
+ List("sha1:00904C3YNNEGH5WAG5ZA9XWAEBNXJWT6", statusType2),
+ List("sha1:249C3YNNEGH5WAG5ZAAXWAEBNXJWT6CZ", statusType1),
+ List("sha1:095893C3YNNEGH5WAG5ZAAXWAEBNXJWT", statusType2)
+ )
+
+ val statusType1Count = sampleData.count(lst => lst(1) == statusType1)
+ val statusType2Count = sampleData.count(lst => lst(1) == statusType2)
+
+ JobTest("sandcrawler.HBaseStatusCountJob")
+ .arg("test", "")
+ .arg("app.conf.path", "app.conf")
+ .arg("output", output)
+ .arg("debug", "true")
+ .source[Tuple](HBaseStatusCountJob.getHBaseSource,
+ sampleData.map(l => new Tuple(l.map(s => {new ImmutableBytesWritable(Bytes.toBytes(s))}):_*)))
+ .sink[Tuple](Tsv(output)) {
+ outputBuffer =>
+ it("should return a 2-element list.") {
+ assert(outputBuffer.size === 2)
+ }
+
+ // Convert List[Tuple] to Map[String, Integer].
+ val counts = outputBuffer.map(t => (t.getString(0), t.getInteger(1))).toMap
+ it("should have the appropriate number of each status type") {
+ assert(counts(statusType1) == statusType1Count)
+ assert(counts(statusType2) == statusType2Count)
+ }
+ }
+ .run
+ .finish
+}