Cache segmented sentences/tokenized words
Opened this issue · 0 comments
Deleted user commented
Currently makes little sense because our segmentation/tokenization algorithms are just two one-liners, but as soon as we get anything better, it might make sense.
The most important case here is termcheck()
which we execute in excess of five times to achieve the same result. dupecheck()
/sentencelengthcheck()
work a bit differently, so might not make as much sense.