You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When loading in a large corpus of documents, I got the error Error: n not less than length(words), which was ultimately traced to assert_that(n < length(words)) in tokenize_ngrams(). I found this was because I had a very short document, but also because the document was all whitespace characters. I wonder if it would make sense to check for empty documents.
Several things to do here:
Error message should be more descriptive.
Show in the vignette how to filter short documents.
The text was updated successfully, but these errors were encountered:
From rOpenSci onboarding:
Several things to do here:
The text was updated successfully, but these errors were encountered: