diff options
| author | Bryan Newbold <bnewbold@archive.org> | 2019-04-12 13:48:27 -0700 | 
|---|---|---|
| committer | Bryan Newbold <bnewbold@archive.org> | 2019-04-12 14:19:29 -0700 | 
| commit | d93ebaa691f8b200a5761850b4533a153cb457ee (patch) | |
| tree | 7cf0077764e9117afb3ccf3f318aac425be7d3ef /scalding | |
| parent | 8ac10ab7fe310df55ab5a66d741ea25c24389418 (diff) | |
| download | sandcrawler-d93ebaa691f8b200a5761850b4533a153cb457ee.tar.gz sandcrawler-d93ebaa691f8b200a5761850b4533a153cb457ee.zip | |
scalding dump-grobid-status-code job
Diffstat (limited to 'scalding')
| -rw-r--r-- | scalding/src/main/scala/sandcrawler/DumpGrobidStatusCodeJob.scala | 34 | 
1 files changed, 34 insertions, 0 deletions
| diff --git a/scalding/src/main/scala/sandcrawler/DumpGrobidStatusCodeJob.scala b/scalding/src/main/scala/sandcrawler/DumpGrobidStatusCodeJob.scala new file mode 100644 index 0000000..42b3464 --- /dev/null +++ b/scalding/src/main/scala/sandcrawler/DumpGrobidStatusCodeJob.scala @@ -0,0 +1,34 @@ +package sandcrawler + +import java.util.Properties + +import cascading.property.AppProps +import cascading.tuple.Fields +import com.twitter.scalding._ +import com.twitter.scalding.typed.TDsl._ +import org.apache.hadoop.hbase.io.ImmutableBytesWritable +import org.apache.hadoop.hbase.util.Bytes +import parallelai.spyglass.base.JobBase +import parallelai.spyglass.hbase.HBaseConstants.SourceMode +import parallelai.spyglass.hbase.HBasePipeConversions +import parallelai.spyglass.hbase.HBaseSource + +// Dumps status code for each GROBID-processed file. Good for crawl/corpus +// analytics, if we consider GROBID status a rough "is this a paper" metric. +class DumpGrobidStatusCodeJob(args: Args) extends JobBase(args) with HBasePipeConversions { + +  val metaPipe : TypedPipe[(String, Long)] = HBaseBuilder.build(args("hbase-table"), +                     args("zookeeper-hosts"), +                     List("grobid0:status_code"), +                     SourceMode.SCAN_ALL) +    .read +    .toTypedPipe[(ImmutableBytesWritable,ImmutableBytesWritable)](new Fields("key", "status_code")) +    .filter { case (_, status_code) => status_code != null } +    .map { case (key, status_code) => +      (Bytes.toString(key.copyBytes()), +       Bytes.toLong(status_code.copyBytes())) +    }; + +  metaPipe.write(TypedTsv[(String,Long)](args("output"))) + +} | 
