1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
|
package example
import org.junit.runner.RunWith
import com.twitter.scalding.{JobTest, TupleConversions}
import org.scalatest.FunSpec
import org.scalatest.junit.JUnitRunner
import org.slf4j.LoggerFactory
import org.apache.hadoop.hbase.io.ImmutableBytesWritable
import cascading.tuple.{Tuple, Fields}
import org.apache.hadoop.hbase.util.Bytes
import scala._
import com.twitter.scalding.Tsv
import parallelai.spyglass.hbase.HBaseSource
import parallelai.spyglass.hbase.HBaseConstants.SourceMode
/**
* Example of how to define tests for HBaseSource
*/
@RunWith(classOf[JUnitRunner])
class SimpleHBaseSourceExampleTest extends FunSpec with TupleConversions {
val output = "/tmp/testOutput"
val log = LoggerFactory.getLogger(this.getClass.getName)
val sampleData = List(
List("1", "kk1", "pp1"),
List("2", "kk2", "pp2"),
List("3", "kk3", "pp3")
)
JobTest("example.SimpleHBaseSourceExample")
.arg("test", "")
.arg("app.conf.path", "app.conf")
.arg("output", output)
.arg("debug", "true")
.source[Tuple](
new HBaseSource(
"table_name",
"mtrcs-zk1.us.archive.org:2181",
new Fields("key"),
List("column_family"),
List(new Fields("column_name1", "column_name2")),
sourceMode = SourceMode.GET_LIST, keyList = List("1", "2", "3")),
sampleData.map(l => new Tuple(l.map(s => {new ImmutableBytesWritable(Bytes.toBytes(s))}):_*)))
.sink[Tuple](Tsv(output format "get_list")) {
outputBuffer =>
log.debug("Output => " + outputBuffer)
it("should return the test data provided.") {
println("outputBuffer.size => " + outputBuffer.size)
assert(outputBuffer.size === 3)
}
}
.run
.finish
}
|