aboutsummaryrefslogtreecommitdiffstats
path: root/sandcrawler-rfc.md
diff options
context:
space:
mode:
Diffstat (limited to 'sandcrawler-rfc.md')
-rw-r--r--sandcrawler-rfc.md2
1 files changed, 1 insertions, 1 deletions
diff --git a/sandcrawler-rfc.md b/sandcrawler-rfc.md
index fea6a7c..ecf7ab8 100644
--- a/sandcrawler-rfc.md
+++ b/sandcrawler-rfc.md
@@ -73,7 +73,7 @@ process HTML and look for PDF outlinks, but wouldn't crawl recursively.
HBase is used for de-dupe, with records (pointers) stored in WARCs.
A second config would take seeds as entire journal websites, and would crawl
-continously.
+continuously.
Other components of the system "push" tasks to the crawlers by copying schedule
files into the crawl action directories.