diff options
author | bnewbold <bnewbold@archive.org> | 2018-06-07 20:02:42 +0000 |
---|---|---|
committer | bnewbold <bnewbold@archive.org> | 2018-06-07 20:02:42 +0000 |
commit | 410c48faf2099de74292e8583fcd2524d6fd1b7c (patch) | |
tree | 9684b67ee0b2ac8e7ca3a047fc8368477faf19f5 /scalding/src | |
parent | 625ef34f957f7f5fdad99c6ce9d84cf7891fbdef (diff) | |
parent | 6eca6290aa3fc829f4767023ae075350a0a78192 (diff) | |
download | sandcrawler-410c48faf2099de74292e8583fcd2524d6fd1b7c.tar.gz sandcrawler-410c48faf2099de74292e8583fcd2524d6fd1b7c.zip |
Merge branch 'groupby' into 'master'
Added HBaseMimeCount{Job,Test}, which counts the number of rows with each mimetype, as an example of groupby
See merge request webgroup/sandcrawler!4
Diffstat (limited to 'scalding/src')
4 files changed, 178 insertions, 0 deletions
diff --git a/scalding/src/main/scala/sandcrawler/HBaseMimeCountJob.scala b/scalding/src/main/scala/sandcrawler/HBaseMimeCountJob.scala new file mode 100644 index 0000000..819a652 --- /dev/null +++ b/scalding/src/main/scala/sandcrawler/HBaseMimeCountJob.scala @@ -0,0 +1,28 @@ +package sandcrawler + +import cascading.property.AppProps +import cascading.tuple.Fields +import com.twitter.scalding._ +import java.util.Properties +import parallelai.spyglass.base.JobBase +import parallelai.spyglass.hbase.{HBaseSource, HBasePipeConversions} +import parallelai.spyglass.hbase.HBaseConstants.SourceMode + +class HBaseMimeCountJob(args: Args) extends JobBase(args) with HBasePipeConversions { + val output = args("output") + + HBaseMimeCountJob.getHBaseSource + .read + .fromBytesWritable(List('mime)) + .debug + .groupBy('mime){group => group.size('count)} + .write(Tsv(output)) +} + +object HBaseMimeCountJob { + def getHBaseSource = HBaseBuilder.build( + "wbgrp-journal-extract-0-qa", // HBase Table Name + "mtrcs-zk1.us.archive.org:2181", // HBase Zookeeper server (to get runtime config info; can be array?) + List("file:mime"), + SourceMode.SCAN_ALL) +} diff --git a/scalding/src/main/scala/sandcrawler/HBaseStatusCountJob.scala b/scalding/src/main/scala/sandcrawler/HBaseStatusCountJob.scala new file mode 100644 index 0000000..0675efc --- /dev/null +++ b/scalding/src/main/scala/sandcrawler/HBaseStatusCountJob.scala @@ -0,0 +1,28 @@ +package sandcrawler + +import cascading.property.AppProps +import cascading.tuple.Fields +import com.twitter.scalding._ +import java.util.Properties +import parallelai.spyglass.base.JobBase +import parallelai.spyglass.hbase.{HBaseSource, HBasePipeConversions} +import parallelai.spyglass.hbase.HBaseConstants.SourceMode + +class HBaseStatusCountJob(args: Args) extends JobBase(args) with HBasePipeConversions { + val output = args("output") + + HBaseStatusCountJob.getHBaseSource + .read + .fromBytesWritable(List('status)) +// .debug + .groupBy('status){group => group.size('count)} + .write(Tsv(output)) +} + +object HBaseStatusCountJob { + def getHBaseSource = HBaseBuilder.build( + "wbgrp-journal-extract-0-qa", // HBase Table Name + "mtrcs-zk1.us.archive.org:2181", // HBase Zookeeper server (to get runtime config info; can be array?) + List("grobid0:status"), + SourceMode.SCAN_ALL) +} diff --git a/scalding/src/test/scala/sandcrawler/HBaseMimeCountTest.scala b/scalding/src/test/scala/sandcrawler/HBaseMimeCountTest.scala new file mode 100644 index 0000000..eb6f4ff --- /dev/null +++ b/scalding/src/test/scala/sandcrawler/HBaseMimeCountTest.scala @@ -0,0 +1,61 @@ +package sandcrawler + +import cascading.tuple.{Tuple, Fields} +import com.twitter.scalding.{JobTest, Tsv, TupleConversions} +import org.apache.hadoop.hbase.io.ImmutableBytesWritable +import org.apache.hadoop.hbase.util.Bytes +import org.junit.runner.RunWith +import org.scalatest.FunSpec +import org.scalatest.junit.JUnitRunner +import org.slf4j.LoggerFactory +import parallelai.spyglass.hbase.HBaseSource +import parallelai.spyglass.hbase.HBaseConstants.SourceMode +import scala._ + +@RunWith(classOf[JUnitRunner]) +class HBaseMimeCountTest extends FunSpec with TupleConversions { + + val output = "/tmp/testOutput" + + val log = LoggerFactory.getLogger(this.getClass.getName) + + val mimeType1 = "text/html" + val mimeType2 = "application/pdf" + + val sampleData = List( + List("sha1:K2DKSSVTXWPRMFDTWSTCQW3RVWRIOV3Q", mimeType1), + List("sha1:C3YNNEGH5WAG5ZAAXWAEBNXJWT6CZ3WU", mimeType1), + List("sha1:SDKUVHC3YNNEGH5WAG5ZAAXWAEBNX4WT", mimeType2), + List("sha1:35985C3YNNEGH5WAG5ZAAXWAEBNXJW56", mimeType2), + List("sha1:885C3YNNEGH5WAG5ZAAXWA8BNXJWT6CZ", mimeType2), + List("sha1:00904C3YNNEGH5WAG5ZA9XWAEBNXJWT6", mimeType2), + List("sha1:249C3YNNEGH5WAG5ZAAXWAEBNXJWT6CZ", mimeType1), + List("sha1:095893C3YNNEGH5WAG5ZAAXWAEBNXJWT", mimeType2) + ) + + val mimeType1Count = sampleData.count(lst => lst(1) == mimeType1) + val mimeType2Count = sampleData.count(lst => lst(1) == mimeType2) + + JobTest("sandcrawler.HBaseMimeCountJob") + .arg("test", "") + .arg("app.conf.path", "app.conf") + .arg("output", output) + .arg("debug", "true") + .source[Tuple](HBaseMimeCountJob.getHBaseSource, + sampleData.map(l => new Tuple(l.map(s => {new ImmutableBytesWritable(Bytes.toBytes(s))}):_*))) + .sink[Tuple](Tsv(output)) { + outputBuffer => + it("should return a 2-element list.") { + assert(outputBuffer.size === 2) + } + + // Convert List[Tuple] to Map[String, Integer]. + val counts = outputBuffer.map(t => (t.getString(0), t.getInteger(1))).toMap + it("should have the appropriate number of each mime type") { + assert(counts(mimeType1) == mimeType1Count) + assert(counts(mimeType2) == mimeType2Count) + } + } + .run + .finish +} diff --git a/scalding/src/test/scala/sandcrawler/HBaseStatusCountTest.scala b/scalding/src/test/scala/sandcrawler/HBaseStatusCountTest.scala new file mode 100644 index 0000000..8b5c3d6 --- /dev/null +++ b/scalding/src/test/scala/sandcrawler/HBaseStatusCountTest.scala @@ -0,0 +1,61 @@ +package sandcrawler + +import cascading.tuple.{Tuple, Fields} +import com.twitter.scalding.{JobTest, Tsv, TupleConversions} +import org.apache.hadoop.hbase.io.ImmutableBytesWritable +import org.apache.hadoop.hbase.util.Bytes +import org.junit.runner.RunWith +import org.scalatest.FunSpec +import org.scalatest.junit.JUnitRunner +import org.slf4j.LoggerFactory +import parallelai.spyglass.hbase.HBaseSource +import parallelai.spyglass.hbase.HBaseConstants.SourceMode +import scala._ + +@RunWith(classOf[JUnitRunner]) +class HBaseStatusCountTest extends FunSpec with TupleConversions { + + val output = "/tmp/testOutput" + + val log = LoggerFactory.getLogger(this.getClass.getName) + + val statusType1 = "200" + val statusType2 = "404" + + val sampleData = List( + List("sha1:K2DKSSVTXWPRMFDTWSTCQW3RVWRIOV3Q", statusType1), + List("sha1:C3YNNEGH5WAG5ZAAXWAEBNXJWT6CZ3WU", statusType1), + List("sha1:SDKUVHC3YNNEGH5WAG5ZAAXWAEBNX4WT", statusType2), + List("sha1:35985C3YNNEGH5WAG5ZAAXWAEBNXJW56", statusType2), + List("sha1:885C3YNNEGH5WAG5ZAAXWA8BNXJWT6CZ", statusType2), + List("sha1:00904C3YNNEGH5WAG5ZA9XWAEBNXJWT6", statusType2), + List("sha1:249C3YNNEGH5WAG5ZAAXWAEBNXJWT6CZ", statusType1), + List("sha1:095893C3YNNEGH5WAG5ZAAXWAEBNXJWT", statusType2) + ) + + val statusType1Count = sampleData.count(lst => lst(1) == statusType1) + val statusType2Count = sampleData.count(lst => lst(1) == statusType2) + + JobTest("sandcrawler.HBaseStatusCountJob") + .arg("test", "") + .arg("app.conf.path", "app.conf") + .arg("output", output) + .arg("debug", "true") + .source[Tuple](HBaseStatusCountJob.getHBaseSource, + sampleData.map(l => new Tuple(l.map(s => {new ImmutableBytesWritable(Bytes.toBytes(s))}):_*))) + .sink[Tuple](Tsv(output)) { + outputBuffer => + it("should return a 2-element list.") { + assert(outputBuffer.size === 2) + } + + // Convert List[Tuple] to Map[String, Integer]. + val counts = outputBuffer.map(t => (t.getString(0), t.getInteger(1))).toMap + it("should have the appropriate number of each status type") { + assert(counts(statusType1) == statusType1Count) + assert(counts(statusType2) == statusType2Count) + } + } + .run + .finish +} |