aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorEllen Spertus <ellen.spertus@gmail.com>2018-07-19 16:03:54 -0700
committerEllen Spertus <ellen.spertus@gmail.com>2018-07-19 16:03:54 -0700
commit687693e96f5a55aa6d38450da0d10e95df2d9422 (patch)
tree04bb6c63aaccef2da813b19a0ee3c56f3c338a57
parent97c8bdc1973f6dd54216a6f1b0b49dad0660b874 (diff)
downloadsandcrawler-687693e96f5a55aa6d38450da0d10e95df2d9422.tar.gz
sandcrawler-687693e96f5a55aa6d38450da0d10e95df2d9422.zip
Undid changes that did not belong in this branch.
-rw-r--r--scalding/scalding-debugging.md14
-rw-r--r--scalding/src/main/scala/sandcrawler/CdxBackfillJob.scala173
-rw-r--r--scalding/src/test/scala/sandcrawler/CdxBackfillJob.scala117
3 files changed, 0 insertions, 304 deletions
diff --git a/scalding/scalding-debugging.md b/scalding/scalding-debugging.md
index 404fb4d..9143698 100644
--- a/scalding/scalding-debugging.md
+++ b/scalding/scalding-debugging.md
@@ -45,20 +45,6 @@ resolved by ensuring that the `HBaseSource` constructors had exactly identical
names and arguments (eg, table names and zookeeper quorums have to be exact
matches).
-If you get:
-
- value toTypedPipe is not a member of cascading.pipe.Pipe
-
-You probably need to:
-
- import com.twitter.scalding.typed.TDsl._
-
-## Running Individual Tests
-
-You can run a single test matching a string glob pattern like:
-
- sbt:sandcrawler> testOnly *CdxBackfill*
-
## Fields
Values of type `List[Fields]` are not printed in the expected way:
diff --git a/scalding/src/main/scala/sandcrawler/CdxBackfillJob.scala b/scalding/src/main/scala/sandcrawler/CdxBackfillJob.scala
deleted file mode 100644
index 0251e07..0000000
--- a/scalding/src/main/scala/sandcrawler/CdxBackfillJob.scala
+++ /dev/null
@@ -1,173 +0,0 @@
-package sandcrawler
-
-import cascading.property.AppProps
-import cascading.tuple.Fields
-import cascading.pipe.joiner._
-import com.twitter.scalding._
-import com.twitter.scalding.typed.TDsl._
-import java.util.Properties
-import cascading.tap.SinkMode
-import parallelai.spyglass.base.JobBase
-import parallelai.spyglass.hbase.HBaseConstants.SourceMode
-import parallelai.spyglass.hbase.{HBaseSource, HBasePipeConversions}
-
-// Type that represents a raw parsed CDX line
-case class CdxLine(surt: String,
- datetime: String,
- url: String,
- mime: String,
- http_status: String,
- sha1: String,
- c_size: String,
- offset: String,
- warc: String)
-
-
-/**
- * CDX backfill:
- * 1. parse CDX (all columns)
- * 2. filter CDX (pdf, HTTP 200, etc)
- * 3. source HBase (key column only)
- * 4. left join CDX to HBase
- * 5. filter to only those with null HBase key column
- * 6. convert CDX fields to HBase columns
- * 7. sink results to HBase
- *
- * TODO: I really mixed the Scalding "field-base" and "type-based" APIs here.
- * Should decide on a best practice.
- */
-class CdxBackfillJob(args: Args) extends JobBase(args) with HBasePipeConversions {
-
- import CdxBackfillJob._
-
- val hbaseSource = getHBaseSource(args("hbase-table"), args("zookeeper-hosts"))
- val hbaseSink = getHBaseSink(args("hbase-table"), args("zookeeper-hosts"))
-
- // Parse CDX lines from text file to typed pipe
- val lines : TypedPipe[String] = TypedPipe.from(TextLine(args("cdx-input-path")))
- val cdxLines : TypedPipe[CdxLine] = lines
- .filter { isCdxLine }
- .map { lineToCdxLine }
- .filter { CdxBackfillJob.keepCdx(_) }
-
- val cdxRows : TypedPipe[(String, String, String, String)] = cdxLines
- .map { CdxBackfillJob.cdxLineToRow }
-
- val existingKeys : TypedPipe[String] = hbaseSource
- .read
- .toTypedPipe[String]('key)
-
- // filters out all the lines that have an existing SHA1 key in HBase
- // the groupBy statements are to select key values to join on
- val newRows : TypedPipe[(String, String, String, String)] = existingKeys
- .groupBy( identity )
- .rightJoin(cdxRows.groupBy(_._1))
- .toTypedPipe
- .debug
- .collect { case (_, (None, row)) => row }
- .debug
-
- // convert to tuple form and write out into HBase
- newRows
- .toPipe('key, 'c, 'cdx, 'mime)
- .toBytesWritable( new Fields("key", "c", "cdx", "mime") )
- .write(hbaseSink)
-
- // XXX:
- //.toPipe("all")
- //.mapTo('all -> ('key, 'c, 'cdx, 'mime)) { x : (String, String, String, String) => x }
-
-}
-
-object CdxBackfillJob {
-
- def getHBaseSource(hbase_table: String, zookeeper_hosts: String) : HBaseSource = {
- return HBaseBuilder.build(
- hbase_table,
- zookeeper_hosts,
- List("file:size"), // not actually needed
- SourceMode.SCAN_ALL)
- }
-
- def getHBaseSink(hbase_table: String, zookeeper_hosts: String) : HBaseSource = {
- return HBaseBuilder.buildSink(
- hbase_table,
- zookeeper_hosts,
- List("f:c", "file:cdx", "file:mime"),
- SinkMode.UPDATE)
- }
-
- def normalizeMime(raw: String) : String = {
-
- val NORMAL_MIME = List("application/pdf",
- "application/postscript",
- "text/html",
- "text/xml")
-
- val lower = raw.toLowerCase()
- NORMAL_MIME.foreach(norm =>
- if (lower.startsWith(norm)) {
- return norm
- }
- )
-
- // Common special cases
- if (lower.startsWith("application/xml")) {
- return "text/xml"
- }
- if (lower.startsWith("application/x-pdf")) {
- return "application/pdf"
- }
- return lower
-
- }
-
- def isCdxLine(line: String) : Boolean = {
- // malformated or non-CDX11 lines
- !(line.startsWith("#") || line.startsWith(" ") || line.startsWith("filedesc") ||
- line.split(" ").size != 11)
- }
-
- def keepCdx(line: CdxLine) : Boolean = {
- // TODO: sha1.isalnum() and c_size.isdigit() and offset.isdigit() and dt.isdigit()
- if (line.http_status != "200" || line.sha1.size != 32) {
- return false
- }
- // TODO: '-' in (line.surt, line.datetime, line.url, line.mime, line.c_size, line.offset, line.warc)
- return true
- }
-
- // Returns (key, f:c, file:cdx, file:mime), all as strings, which is close to
- // how they will be inserted into HBase
- def cdxLineToRow(line: CdxLine) : (String, String, String, String) = {
-
- val key = "sha1:" + line.sha1
-
- val warcFile = line.warc.split('/')(1)
-
- // Read CDX-style datetime and conver to ISO 8601 with second resolution
- val dtFormat = new java.text.SimpleDateFormat("yyyyMMddHHmmss")
- val isoFormat = new java.text.SimpleDateFormat("yyyy-MM-dd'T'HH:mm:ss'Z'")
- // TODO: timezones? UTC to UTC, so I don't think so.
- val dtIso = isoFormat.format(dtFormat.parse(line.datetime))
-
- // warc_file = warc.split('/')[-1]
- // dt_iso = datetime.strptime(dt, "%Y%m%d%H%M%S").isoformat()
- // f:c = dict(u=url, d=dt_iso, f=warc_file, o=int(offset), c=1)
-
- // This is the "f:c" field. 'i' intentionally not set
- val heritrixInfo = ""
-
- // file:cdx = dict(surt=surt, dt=dt, url=url, c_size=int(c_size),
- // offset=int(offset), warc=warc)
- val fileCdx = ""
- (key, heritrixInfo, fileCdx, line.mime)
- }
-
- def lineToCdxLine(line: String) : CdxLine = {
- val raw = line.split("\\s+")
- // surt, datetime, url, mime, http_status, sha1, SKIP, SKIP, c_size, offset, warc
- CdxLine(raw(0), raw(1), raw(2), raw(3), raw(4), raw(5), raw(8), raw(9), raw(10))
- }
-
-}
diff --git a/scalding/src/test/scala/sandcrawler/CdxBackfillJob.scala b/scalding/src/test/scala/sandcrawler/CdxBackfillJob.scala
deleted file mode 100644
index de94494..0000000
--- a/scalding/src/test/scala/sandcrawler/CdxBackfillJob.scala
+++ /dev/null
@@ -1,117 +0,0 @@
-
-package sandcrawler
-
-import org.scalatest._
-import cascading.tuple.{Tuple, Fields}
-import com.twitter.scalding.{JobTest, Tsv, TypedTsv, TupleConversions, TextLine}
-import org.apache.hadoop.hbase.io.ImmutableBytesWritable
-import org.apache.hadoop.hbase.util.Bytes
-import org.junit.runner.RunWith
-import org.scalatest.FunSpec
-import org.scalatest.junit.JUnitRunner
-import org.slf4j.LoggerFactory
-import parallelai.spyglass.hbase.HBaseSource
-import parallelai.spyglass.hbase.HBaseConstants.SourceMode
-
-class CdxBackfillTest extends FlatSpec with Matchers {
-
- import CdxBackfillJob._
-
- it should "normalize mimetypes" in {
- assert(CdxBackfillJob.normalizeMime("asdf") === "asdf")
- assert(CdxBackfillJob.normalizeMime("application/pdf") === "application/pdf")
- assert(CdxBackfillJob.normalizeMime("application/pdf+journal") === "application/pdf")
- assert(CdxBackfillJob.normalizeMime("Application/PDF") === "application/pdf")
- assert(CdxBackfillJob.normalizeMime("application/p") === "application/p")
- assert(CdxBackfillJob.normalizeMime("application/xml+stuff") === "text/xml")
- assert(CdxBackfillJob.normalizeMime("application/x-pdf") === "application/pdf")
- assert(CdxBackfillJob.normalizeMime("application/x-html") === "application/x-html")
- }
-
- it should "filter CDX lines" in {
- assert(true === keepCdx(lineToCdxLine(
- "edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz")))
- // redirect
- assert(false === keepCdx(lineToCdxLine(
- "edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 301 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz")))
- }
-
- it should "know what CDX lines are" in {
- assert(true === isCdxLine(
- "edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz"))
- assert(false === isCdxLine(""))
- assert(false === isCdxLine(
- " edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz"))
- assert(false === isCdxLine(
- "#edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz"))
- // missing two fields
- assert(false === isCdxLine(
- "edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz"))
- // extra field
- assert(false === isCdxLine(
- "edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz -"))
- }
-
- it should "execute lineToRow" in {
- cdxLineToRow(lineToCdxLine(
- "edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz"))
- }
-
-}
-
-@RunWith(classOf[JUnitRunner])
-class CdxBackfillJobTest extends FunSpec with TupleConversions {
-
- val output = "/tmp/testOutput"
- val (testTable, testHost, testCdxFile) = ("test-table", "dummy-host:2181", "test_file.cdx")
-
- val log = LoggerFactory.getLogger(this.getClass.getName)
-
- val dummySizeBytes = Bytes.toBytes(100)
-
- val sampleData = List(
- List(Bytes.toBytes("sha1:K2DKSSVTXWPRMFDTWSTCQW3RVWRIOV3Q"), dummySizeBytes),
- List(Bytes.toBytes("sha1:C3YNNEGH5WAG5ZAAXWAEBNXJWT6CZ3WU"), dummySizeBytes),
- List(Bytes.toBytes("sha1:SDKUVHC3YNNEGH5WAG5ZAAXWAEBNX4WT"), dummySizeBytes),
- List(Bytes.toBytes("sha1:095893C3YNNEGH5WAG5ZAAXWAEBNXJWT"), dummySizeBytes)
- )
- val sampleCdxLines = List(
- // clean line
- "edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz",
- // has existing SHA1
- "edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 C3YNNEGH5WAG5ZAAXWAEBNXJWT6CZ3WU - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz",
- // HTTP status code
- "edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 301 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz",
- // not CDX (prefixed with hash)
- "#edu,upenn,ldc)/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf 20170828233154 https://www.ldc.upenn.edu/sites/www.ldc.upenn.edu/files/medar2009-large-arabic-broadcast-collection.pdf application/pdf 200 WL3FEA62TEU4F52Y5DOVQ62VET4QJW7G - - 210251 931661233 SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828231135742-00000-00009-wbgrp-svc284/SEMSCHOLAR-PDF-CRAWL-2017-08-04-20170828232253025-00005-3480~wbgrp-svc284.us.archive.org~8443.warc.gz"
- )
-
- JobTest("sandcrawler.CdxBackfillJob")
- .arg("test", "")
- .arg("app.conf.path", "app.conf")
- .arg("output", output)
- .arg("hbase-table", testTable)
- .arg("zookeeper-hosts", testHost)
- .arg("cdx-input-path", testCdxFile)
- .arg("debug", "true")
- .source[Tuple](CdxBackfillJob.getHBaseSource(testTable, testHost),
- sampleData.map(l => new Tuple(l.map(s => {new ImmutableBytesWritable(s)}):_*)))
- .source[String](TextLine(testCdxFile), sampleCdxLines)
- .sink[Tuple](CdxBackfillJob.getHBaseSink(testTable, testHost)) {
- outputBuffer =>
-
- it("should return a 1-element list (after join).") {
- // XXX:
- assert(outputBuffer.size === 1)
- }
-
- // Convert List[Tuple] to Map[Long, Long].
- val counts = outputBuffer.map(t => (t.getLong(0), t.getLong(1))).toMap
- it("should have the appropriate number of each status type") {
- // XXX:
- assert(counts(1) == 3)
- }
- }
- .run
- .finish
-}