diff options
author | Martin Czygan <martin.czygan@gmail.com> | 2020-12-18 02:59:48 +0100 |
---|---|---|
committer | Martin Czygan <martin.czygan@gmail.com> | 2020-12-18 02:59:48 +0100 |
commit | 233411895be924c68c50a95f52269d49290a6a14 (patch) | |
tree | 986710be4e26b9a9db44adec738f03c51d446ced | |
parent | e463ff2e949bd02e5fe70be67f2736dc09d59960 (diff) | |
download | fuzzycat-233411895be924c68c50a95f52269d49290a6a14.tar.gz fuzzycat-233411895be924c68c50a95f52269d49290a6a14.zip |
update README
-rw-r--r-- | README.md | 44 | ||||
-rw-r--r-- | notes/steps.dot | 14 | ||||
-rw-r--r-- | notes/steps.png | bin | 0 -> 29435 bytes |
3 files changed, 47 insertions, 11 deletions
@@ -1,24 +1,46 @@ # fuzzycat (wip) -Fuzzy matching publications for [fatcat](https://fatcat.wiki). +Fuzzy matching utilities for [fatcat](https://fatcat.wiki). ![https://pypi.org/project/fuzzycat/](https://img.shields.io/pypi/v/fuzzycat?style=flat-square) -# Example Run +## Dataset -Run any clustering algorithm. +For development, we worked on a `release_export_expanded.json` dump (113G/700G +zstd/plain, XXX lines) and with the [fatcat API](https://api.fatcat.wiki/). + +Workflow Fall 2020 + +![](notes/steps.png) + +## Facilities + +### Clustering + +Derive cluster of similar documents from a [fatcat database release +dump](https://archive.org/details/fatcat_snapshots_and_exports?&sort=-publicdate). + +Following algorithms are implemented (or planned): + +* [x] exact title matches (title) +* [x] normalized title matches (tnorm) +* [x] NYSIIS encoded title matches (tnysi) +* [x] extended title normalization (tsandcrawler) + +Example running clustering: ``` -$ time python -m fuzzycat cluster -t tsandcrawler < data/sample10m.json | \ - zstd -c9 > sample_cluster.json.zst -2020-11-18 00:19:48.194 DEBUG __main__ - run_cluster: - {"key_fail": 0, "key_ok": 9999938, "key_empty": 62, "key_denylist": 0, "num_clusters": 9040789} - -real 75m23.045s -user 95m14.455s -sys 3m39.121s +$ python -m fuzzycat cluster -t tsandcrawler < data/re.json > cluster.json.zst ``` +Clustering works in a three step process: + +1. key extraction for each document (choose algorithm) +2. sorting by keys (via GNU sort) +3. group by key and write out ([itertools.groupby](https://docs.python.org/3/library/itertools.html#itertools.groupby)) + +### Verification + Run verification. ``` diff --git a/notes/steps.dot b/notes/steps.dot new file mode 100644 index 0000000..1ebb504 --- /dev/null +++ b/notes/steps.dot @@ -0,0 +1,14 @@ +digraph steps { + node [shape=note, fontname="helvetica"]; + graph [fontname = "helvetica"]; + edge [fontname = "helvetica"]; + + "release_export_expanded.json" -> "clustered" [label="fuzzycat cluster"]; + "clustered" -> "verified" [label="fuzzycat verify"]; + "verified" -> "manual review"; + "manual review" -> "discover pattern"; + "manual review" -> "fix codepath"; + "fix codepath" -> "add verification test case"; + "discover pattern" -> "add verification test case"; + "add verification test case" -> "clustered" [label="repeat"]; +} diff --git a/notes/steps.png b/notes/steps.png Binary files differnew file mode 100644 index 0000000..181dacf --- /dev/null +++ b/notes/steps.png |