when using a custom tokenizer, how would you invoke WordsCount.count? first tokenize the text, then re-join the tokens into a string and run .count on it? Or is there a more efficient way?
tk = WordsCounted::Tokeniser.new( doc.force_encoding('UTF-8') ).tokenise(exclude: Stopwords::STOPWORDS)
wc = WordsCounted.count tk.join(' ')